VCS new volume created. Need to add up under existing Resource Group under 2 node clustered Solaris 10 Server...Important..please assist
Hi, I am newbi to VCS. OS is Solaris10. Two node cluster env. New filesystem already created and mounted with dedicated Vertual IP and is already under VCS. We have added two more Vol's (Vol03 & Vol04) in the existing Resource Group say 'node1_rg' Guys, please help me out to know the sequential and step by step procedure with commands to take the newly created Volumes under existing Resource Group......and how to carry out the Failover Test. this is a test server on which we need to do the task. Oracle Database is also there and we can coordinate with App & DBA team to do this activity...... Pleaseeeeeeee help. this is first time I had been to this forum seeking the help... Thanks, Rinku, Pune- IndiaSolved3.3KViews2likes9CommentsUnable to bring the Service Group online.
Hi All, I tried to bring a SG online in a node but it's not comming online. Let me explaing the issue. We did reboot of a node aixprd001 and we found that /etc/filesystem is corrupted so the SG bosinit_SG is in partial state since lot of cluster FS in not mounted. Then we corrected the entry and done the manual mout of all the FS but the SG still show the status partial so we did the bellow command. hagrp -clear bosinit_SG -all Once done the SG is in online state. For safer side we tried to offline the SG and brought it up online again but the SG failed to come online, Bellow is the only error we able find the engine_A.log file. 2014/12/17 06:49:04 VCS NOTICE V-16-1-10166 Initiating manual online of group bosinit_SG on system aixprd001 2014/12/17 06:49:04 VCS NOTICE V-16-1-10233 Clearing Restart attribute for group bosinit_SG on all nodes Please help me by providing suggestion, I will provide the output of logs if needed. Thanks, RufusSolved2KViews0likes4Commentsneed a solution
we have 2 node cluster and with version 5.1 we experienced outage and I think it was due to below error messages can someone shed some light on these messages qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop OFFLINE qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop ONLINE fctl: [ID 999315 kern.warning] WARNING: fctl(4): AL_PA=0xe8 doesn't exist in LILP map scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 243001 kern.info] /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0 (fcp4): offlining lun=0 (trace=0), target=e8 (trace=2800004) vxdmp: [ID 631182 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 removed disk array 600A0B800075F9D9000000004D2334F5, datype = ST2540- vxdmp: [ID 443116 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 i/o error occured (errno=0x6) on dmpnode 334/0x2c last message repeated 59 times vxdmp: [ID 480808 kern.notice] NOTICE: VxVM vxdmp V-5-0-112 disabled path 118/0x18 belonging to the dmpnode 334/0x28 due to open failure vxdmp: [ID 824220 kern.notice] NOTICE: VxVM vxdmp V-5-0-111 disabled dmpnode 334/0x28 what is this dmpnode 334/0x28 signify, I forget how to map this to device as i only remember is tht its in hexadecimal. Also, what could be the cause of it ... is it due to HBA as issue starts with the message like below qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop OFFLINE qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop ONLINE fctl: [ID 999315 kern.warning] WARNING: fctl(4): AL_PA=0xe8 doesn't exist in LILP map1.8KViews2likes4CommentsSG is not switching to next node.
Hi All, I am new to VCS but good in HACMP. In our environment we are using VCS-6.0, I one server we found that the SG is not moving from one node to another node when we tried manual failover using the bellow command. hagrp -switch <SGnamg> -to <sysname> We able to see that the SG is offline in the currnent node but it's not coming online in the secondary node. There is no error locked in engine_A.log except the bellow entry cpus load more than 60% <Secondary node name> Can anyone help me to find the solution for this. I will provide the output of any commands if you need more info to help me out to get this trouble shooted :) Thanks,Solved1.8KViews1like8CommentsSQL memory management in active/active configuration
Hi, I willhave 4 nodes in an active/active/active/active configuration. Each node will have 3 SQL instances installed on it. Each node has 256GB RAM. I know I can set the limit each instance can use, but the more memory SQL gets, the better it runs. Ideally, I would set each instance to use 80GB or so. Roughly 1/3 of the node's memory, leaving some for the OS and Veritas. What happens if one of the nodes goes down? Where would the newly failed over instance get it's memory from? Is there a way to manage this? Should I let SQL manage it? I don't want a run-away query on one instance to hog all the node's memory though and affect the other instances on that node. What are my options?Solved1.6KViews0likes2CommentsCFS Health Check ?
Hi All, We have recently built a 4-node Oracle single-instance CFS cluster, with more to on the way. The cluster will host many databases, and all nodes will be active. The databases will migrate from non-clustered to CFS clustered servers, and we are concerned about what effect CFS may have on performance. Concerns: - Suitability of our LLT links to handle intra-node communication generated by CFS ? - Location of CFS primary, for each filesytem, and how to determine if it is optimal ? - When multiple nodes write to the same filesystem simultaeously, would that be detrimental to performance ? - I am aware of GLM lock/release as one source of CFS intra-node traffic, but wondering if there are other significant sources, such as VxFS cache coherency ? - How to best monitor CFS overhead, intra-node traffic, and overall health ? Thank your for reading ! Ken W ---------------------Configuration--------------------- Red Hat RHEL 6.5 SFCFSHA 6.1.1 with P200 patch 2 x 1 Gbit ethernet as LLT links, on dedicated VLANs Oracle 11.2.0.3 and 12.1.0.2 Oracle code local to each node (outside of cluster control) Oracle data files on CFS filesystem dedicated to only datases running on that node. Oracle Redo on a single shared filesytem for all nodes. -------------------------------------------------------------Solved1.4KViews0likes3CommentsDisk space reporting
I'm runing VCS 5.0 and have a SG with NFS shares off a RAID from the two servers in my cluster. The problem I'm running into is a conflict ofhow muchdisk space is being used with a"df" compared to a "du". One of the file systems is off by 36GB. We've unshared the file system, took down the cluster, and tried an fsck, but no errors were reported and no changes to space being reported as a result. Has anyone ever seen this before?Solved1.1KViews0likes4CommentsVMotion vs VCS
We have a slightly complicated question, it is, however, not really that complicated. We are moving one of our money generating apps over to dedicated UCS blades, and in VCS, (Linux on UCS blades with VCS), there is, however, some talk of using VMware and vMotion, and treat this application like other VM's that we have - in the sense, vMotion them in case of failover. This app has Terabytes of data. It is currently running on powerful Solaris systems and will be migrated to Linux on UCS. I sense that VMware is good for small databases and small apps..... My thinking is that changes should occur at higher levels, thereby insulating lower levels, given this logic, for instance, a failover should occur at the VCS level, not below the os....things like, if the os gets corrupted, vMotion will move the corrupted data alongwith. What do you recommend and why?1.1KViews0likes1Comment2-node VCS CFS - du command takes 10x longer on one node!
Hi All, I have setup a 2 node VCS/CFS Cluster. Both nodes are identical HP Blades (460G6 I think), storage is on EMC Symmetrix arrays. Servers are split between data centers and communications is over Sonet Ring with Storage going over DWDM. Both nodes are running RedHat 5u4 64-bit and have identical packages and kernel parameters. Both were built from the same kickstart files. 2 QLogic HBA's are in each blade and DMP is managing the multipathing. Storage is divided into 2 plexes mirrored by VxVM over the data centers. One plex is on an array(s) in each data center. iozone profiling shows I/O rates are pretty identical from both nodes. du -hsc on shared filesystem is the only issue I'm currently seeing. The shared filesystem has 66,000 files in it currently. du on one node takes about 3 minutes, but on the other node takes anywhere from 30-48 minutes. No other activity is occuring. I've tried changing the master, but it doesn't seem to matter. I have strace output if that would help, but my analysis shows that lstat on the slow node is much slower (.001 on ave) so that times 66,000 kind of gets close to the time disparity. Also noted the an lstat immediately following getdent is much, much slower than subsequent lstat commands. Has anyone seen this behavior? Thanks, Danté978Views0likes4CommentsInformation on the CPU/Memory overhead of using SFWHA-DR versus native Windows
Hi everyone, We are currenlty having a bake-off for Exchange 2010 of using SFWHA 6.0.1 or native a native Microsoft DAG for a COOP/DR solution at he customer site I support. My question for Symatec would be if there is any documentation on the amount of extra overhead on each Exchange 2010 mailbox server when using SFWHA versus just the Windows 2008 operating system itself? The plans are for: 6 Mail Servers 4 Databases per server Windows 2008 R2 with each virtual server with 64 GB RAM If there ae any planning guides from Symantec for Exchange 2010 using SFWHA that would be useful as well. Thanks in advance, ChipSolved920Views0likes2Comments