Disk space reporting
I'm runing VCS 5.0 and have a SG with NFS shares off a RAID from the two servers in my cluster. The problem I'm running into is a conflict ofhow muchdisk space is being used with a"df" compared to a "du". One of the file systems is off by 36GB. We've unshared the file system, took down the cluster, and tried an fsck, but no errors were reported and no changes to space being reported as a result. Has anyone ever seen this before?Solved1.1KViews0likes4CommentsCFS Health Check ?
Hi All, We have recently built a 4-node Oracle single-instance CFS cluster, with more to on the way. The cluster will host many databases, and all nodes will be active. The databases will migrate from non-clustered to CFS clustered servers, and we are concerned about what effect CFS may have on performance. Concerns: - Suitability of our LLT links to handle intra-node communication generated by CFS ? - Location of CFS primary, for each filesytem, and how to determine if it is optimal ? - When multiple nodes write to the same filesystem simultaeously, would that be detrimental to performance ? - I am aware of GLM lock/release as one source of CFS intra-node traffic, but wondering if there are other significant sources, such as VxFS cache coherency ? - How to best monitor CFS overhead, intra-node traffic, and overall health ? Thank your for reading ! Ken W ---------------------Configuration--------------------- Red Hat RHEL 6.5 SFCFSHA 6.1.1 with P200 patch 2 x 1 Gbit ethernet as LLT links, on dedicated VLANs Oracle 11.2.0.3 and 12.1.0.2 Oracle code local to each node (outside of cluster control) Oracle data files on CFS filesystem dedicated to only datases running on that node. Oracle Redo on a single shared filesytem for all nodes. -------------------------------------------------------------Solved1.4KViews0likes3CommentsUnable to bring the Service Group online.
Hi All, I tried to bring a SG online in a node but it's not comming online. Let me explaing the issue. We did reboot of a node aixprd001 and we found that /etc/filesystem is corrupted so the SG bosinit_SG is in partial state since lot of cluster FS in not mounted. Then we corrected the entry and done the manual mout of all the FS but the SG still show the status partial so we did the bellow command. hagrp -clear bosinit_SG -all Once done the SG is in online state. For safer side we tried to offline the SG and brought it up online again but the SG failed to come online, Bellow is the only error we able find the engine_A.log file. 2014/12/17 06:49:04 VCS NOTICE V-16-1-10166 Initiating manual online of group bosinit_SG on system aixprd001 2014/12/17 06:49:04 VCS NOTICE V-16-1-10233 Clearing Restart attribute for group bosinit_SG on all nodes Please help me by providing suggestion, I will provide the output of logs if needed. Thanks, RufusSolved2.1KViews0likes4CommentsSG is not switching to next node.
Hi All, I am new to VCS but good in HACMP. In our environment we are using VCS-6.0, I one server we found that the SG is not moving from one node to another node when we tried manual failover using the bellow command. hagrp -switch <SGnamg> -to <sysname> We able to see that the SG is offline in the currnent node but it's not coming online in the secondary node. There is no error locked in engine_A.log except the bellow entry cpus load more than 60% <Secondary node name> Can anyone help me to find the solution for this. I will provide the output of any commands if you need more info to help me out to get this trouble shooted :) Thanks,Solved1.8KViews1like8CommentsBonding VCS
Hi, I want to reconfigure my clusters so that it runs on bonding mode. The cluster was running only running on specific eth on public facing network. But now since the servers have configured bonding using two eth interfaces, any procedure to follow to change the vcs configuration to use the bonding pairs? ThanksSolved772Views1like1CommentVMotion vs VCS
We have a slightly complicated question, it is, however, not really that complicated. We are moving one of our money generating apps over to dedicated UCS blades, and in VCS, (Linux on UCS blades with VCS), there is, however, some talk of using VMware and vMotion, and treat this application like other VM's that we have - in the sense, vMotion them in case of failover. This app has Terabytes of data. It is currently running on powerful Solaris systems and will be migrated to Linux on UCS. I sense that VMware is good for small databases and small apps..... My thinking is that changes should occur at higher levels, thereby insulating lower levels, given this logic, for instance, a failover should occur at the VCS level, not below the os....things like, if the os gets corrupted, vMotion will move the corrupted data alongwith. What do you recommend and why?1.1KViews0likes1CommentV-16-2-13027 Oracle Rac
Dears , Engine_A is started to be flooded withV-16-2-13027 error code , eventually some resources fail after passing multiple failed monitoringcycle ,clean iscalled . - PrivNIC isconfigured with 2 interfaces , which are the same two interfaces used forcluster inter-communications . - lltstat errors shows : 60 Snd not connected 0 Snd no buffer 0 Snd stream flow drops 1636916 Snd no links up - ps -ef don't show hanged moniroing process or something . - Server performance is fine in general , no recentchanges was done . Any ideas ?861Views1like5Commentsneed a solution
we have 2 node cluster and with version 5.1 we experienced outage and I think it was due to below error messages can someone shed some light on these messages qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop OFFLINE qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop ONLINE fctl: [ID 999315 kern.warning] WARNING: fctl(4): AL_PA=0xe8 doesn't exist in LILP map scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 107833 kern.warning] WARNING: /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0/ssd@w203400a0b875f9d9,0 (ssd3): Command failed to complete...Device is gone scsi: [ID 243001 kern.info] /pci@0,600000/pci@0/pci@9/SUNW,qlc@0/fp@0,0 (fcp4): offlining lun=0 (trace=0), target=e8 (trace=2800004) vxdmp: [ID 631182 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 removed disk array 600A0B800075F9D9000000004D2334F5, datype = ST2540- vxdmp: [ID 443116 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 i/o error occured (errno=0x6) on dmpnode 334/0x2c last message repeated 59 times vxdmp: [ID 480808 kern.notice] NOTICE: VxVM vxdmp V-5-0-112 disabled path 118/0x18 belonging to the dmpnode 334/0x28 due to open failure vxdmp: [ID 824220 kern.notice] NOTICE: VxVM vxdmp V-5-0-111 disabled dmpnode 334/0x28 what is this dmpnode 334/0x28 signify, I forget how to map this to device as i only remember is tht its in hexadecimal. Also, what could be the cause of it ... is it due to HBA as issue starts with the message like below qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop OFFLINE qlc: [ID 630585 kern.info] NOTICE: Qlogic qlc(1): Loop ONLINE fctl: [ID 999315 kern.warning] WARNING: fctl(4): AL_PA=0xe8 doesn't exist in LILP map1.8KViews2likes4CommentsLoadbalacing
HI, I have 2 VMs that runs Server 2008 R2 Standard. I have mapped a LUN as RDM to one of the server. Is there anyway i can loadbalance or share the same drive between the servers at the same time using VRTS STORAGE FOUNDATION HA,/DR 5.1 WIN FOR OS TIER ENTERPRISE EDITION STD LIC EXPRESS BAND S thanks Faisal SP815Views1like4Comments