Veritas cluster issue
Hi We have a VCS of two node. Two file systems are under VCS configured from VXVM. One file system got 100% full.Now we have rebooted the cluster node & the mount points started to show,but after some time that file system got disappear,Checked with concerned DG that also got disabed,we tried to manually import & got succeded.But after starting the volumes we tried to mount the filesystem under the specified mount point. But the below error giving--- # mount /dev/vx/dsk/bgw1dg/vol01 /var/opt/BGw/Server1 mount: /dev/vx/dsk/bgw1dg/vol01 is not this fstype # Kindly suggest what could be the cause. regards...ArupSolved2.2KViews4likes5Commentsvcs resource status abnormal
I am facing a problem,i found a resource become online monitor timeout , and i don't know how to fix it ; could anybody give me some suggestion. thanks; #hastatus .............. group resource system message --------------- -------------------- -------------------- -------------------- cfsmount1 CCCG09BER ONLINE cfsmount1 CCCG10BER ONLINE cfsmount2 CCCG09BER ONLINE cfsmount2 CCCG10BER ONLINE ------------------------------------------------------------------------- cvmvoldg1 CCCG09BER ONLINE cvmvoldg1 CCCG10BER ONLINE group resource system message --------------- -------------------- --------------- -------------------- cvmvoldg1 CCCG09BER |MONITOR TIMEDOUT|UNABLE TO OFFLINE| cvmvoldg2 CCCG09BER ONLINE cvmvoldg2 CCCG10BER ONLINE cvmvoldg2 CCCG09BER |MONITOR TIMEDOUT|UNABLE TO OFFLINE| --------------------------------------------------------------------------------------------------------------------------------------------- # hares -state cvmvoldg1 #Resource Attribute System Value cvmvoldg1 State CCCG09BER ONLINE|MONITOR TIMEDOUT cvmvoldg1 State CCCG10BER ONLINE # # vxdisk list DEVICE TYPE DISK GROUP STATUS c0t0d0s2 auto:SVM - - SVM c0t1d0s2 auto:SVM - - SVM c0t2d0s2 auto:none - - online invalid c0t3d0s2 auto:none - - online invalid c3t0d0 auto:cdsdisk - - online c3t0d1 auto:sliced mmdatadgemcpower4 mmdatadg online shared c3t0d2 auto:cdsdisk - - online c3t0d3 auto:cdsdisk - - online c3t0d4 auto:sliced mmdbdgemcpower2 mmdbdg online shared c3t0d5s2 auto:sliced mmdatadgc3t0d5 mmdatadg online shared c3t0d6s2 auto:sliced mmdatadgc3t0d6 mmdatadg online shared c3t0d7s2 auto:sliced mmdatadgc3t0d7 mmdatadg online shared c3t0d8s2 auto:sliced mmdatadgc3t0d8 mmdatadg online shared c3t0d9s2 auto:sliced mmdatadgc3t0d9 mmdatadg online shared c3t0d10s2 auto:sliced mmdatadgc3t0d10 mmdatadg online shared # vxprint Disk group: mmdbdg TY NAME ASSOC KSTATE LENGTH PLOFFS STATE TUTIL0 PUTIL0 dg mmdbdg mmdbdg - - - - - - dm mmdbdgemcpower2 c3t0d4 - 419348384 - - - - v vol01 fsgen ENABLED 419346432 - ACTIVE - - pl vol01-01 vol01 ENABLED 419346432 - ACTIVE - - sd mmdbdgemcpower2-01 vol01-01 ENABLED 419346432 0 - - - Disk group: mmdatadg TY NAME ASSOC KSTATE LENGTH PLOFFS STATE TUTIL0 PUTIL0 dg mmdatadg mmdatadg - - - - - - dm mmdatadgc3t0d5 c3t0d5s2 - 1467866928 - - - - dm mmdatadgc3t0d6 c3t0d6s2 - 1887246288 - - - - dm mmdatadgc3t0d7 c3t0d7s2 - 1836919616 - - - - dm mmdatadgc3t0d8 c3t0d8s2 - 1887246288 - - - - dm mmdatadgc3t0d9 c3t0d9s2 - 1887246288 - - - - dm mmdatadgc3t0d10 c3t0d10s2 - 1417511040 - - - - dm mmdatadgemcpower4 c3t0d1 - 1887354784 - - - - v vol01 fsgen ENABLED 12271390720 - ACTIVE - - pl vol01-01 vol01 ENABLED 12271390720 - ACTIVE - - sd mmdatadgemcpower4-01 vol01-01 ENABLED 1887354784 0 - - - sd mmdatadgc3t0d5-01 vol01-01 ENABLED 1467866928 1887354784 - - - sd mmdatadgc3t0d6-01 vol01-01 ENABLED 1887246288 3355221712 - - - sd mmdatadgc3t0d7-01 vol01-01 ENABLED 1836919616 5242468000 - - - sd mmdatadgc3t0d8-01 vol01-01 ENABLED 1887246288 7079387616 - - - sd mmdatadgc3t0d9-01 vol01-01 ENABLED 1887246288 8966633904 - - - sd mmdatadgc3t0d10-01 vol01-01 ENABLED 1417510528 10853880192 - - - messages in engine_A log: 2015/03/06 13:44:21 VCS ERROR V-16-2-13027 (CCCG09BER) Resource(cvmvoldg2) - monitor procedure did not complete within the expected time. 2015/03/06 13:44:22 VCS ERROR V-16-2-13027 (CCCG09BER) Resource(cvmvoldg1) - monitor procedure did not complete within the expected time. 2015/03/06 13:46:28 VCS INFO V-16-1-50086 CPU usage on CCCG09BER is 61% 2015/03/06 13:50:21 VCS ERROR V-16-2-13210 (CCCG09BER) Agent is calling clean for resource(cvmvoldg2) because 4 successive invocatio ns of the monitor procedure did not complete within the expected time. 2015/03/06 13:50:21 VCS ERROR V-16-2-13210 (CCCG09BER) Agent is calling clean for resource(cvmvoldg1) because 4 successive invocatio ns of the monitor procedure did not complete within the expected time. 2015/03/06 13:50:39 VCS INFO V-16-2-13068 (CCCG09BER) Resource(cvmvoldg2) - clean completed successfully. 2015/03/06 13:50:53 VCS INFO V-16-2-13068 (CCCG09BER) Resource(cvmvoldg1) - clean completed successfully. 2015/03/06 13:51:40 VCS ERROR V-16-2-13077 (CCCG09BER) Agent is unable to offline resource(cvmvoldg2). Administrative intervention m ay be required. 2015/03/06 13:51:55 VCS ERROR V-16-2-13077 (CCCG09BER) Agent is unable to offline resource(cvmvoldg1). Administrative intervention m ay be required.Solved1.1KViews3likes1Commentgco fencing
hi, i think that vxfencingcould besetup only within evey cluster from every site.i am not aware if vxfencing could be setup as global fencing.suppose i have a node in site 1 and other node in site 2,and a failover group uses node1 or node2.if the heartbeat link between the node1 and node2 fails,then a split brain happens.then how this issue is solved not to corrupt data? in my opinion the advantage of gco over classic disaster recovery is the speed of recovery,i mean the failover groups will failover fast on the other site if a site goes down.right? tnx a lot.Solved1.6KViews3likes6Commentsvcs less nodes
hi, a vcs could runs with less nodes then in normal configuration?i.e gabconfig seeds this vcs with 6 nodes,but in time some of the nodes left the vcs,so now we have 2 nodes in vcs.I am wondering why Symantec decided to have vcs without disk quorum like solaris cluster,safeguard,so on?Solved1KViews3likes4CommentsResources offline again and again
Hi Team, Yesterday, I had an issue as resouces are coming to offline state again and again after manually making it online.I didn't have an output now, however just want to know the reason behind it and what could be its resolution. OS: Sol 10 VCS 6.0 Help would be appreciated..Thanks..Solved775Views3likes2CommentsVCS 5.1 - DG import fails to the new node
Hi, I have recently added a new node to a cluster and it joined the cluster with no problem... but whe I tried to import a DG to the node it fails... Note that this DG is available and imported in a different node. Here si the log when it WORKED in a existing node: 2014/01/16 00:49:42 VCS NOTICE V-16-1-10301 Initiating Online of Resource md_DG (Owner: unknown, Group: mdm) on System D-node6 2014/01/16 00:49:44 VCS WARNING V-16-10001-1014 (D-node6) DiskGroup:md_DG:online:Diskgroups will be imported without reservations 2014/01/16 00:49:46 VCS NOTICE V-16-10001-1009 (D-node6) DiskGroup:md_DG:online:vxdg import succeeded on Disk Group md_DG 2014/01/16 00:49:47 VCS NOTICE V-16-10001-1010 (D-node6) DiskGroup:md_DG:online:Volumes in Disk Group md_DG are started. Any mirrors are updated in background And here is when it FAILED in the new node: 2014/01/16 00:45:11 VCS NOTICE V-16-1-10301 Initiating Online of Resource md_DG (Owner: unknown, Group: mdm) on System d-node9 2014/01/16 00:45:12 VCS WARNING V-16-10001-1014 (d-node9) DiskGroup:md_DG:online:Diskgroups will be imported without reservations 2014/01/16 00:45:12 VCS WARNING V-16-10001-1016 (d-node9) DiskGroup:md_DG:online:vxdg import (clear flag) failed. 2014/01/16 00:45:12 VCS WARNING V-16-10001-1017 (d-node9) DiskGroup:md_DG:online:Trying force import for the diskgroup. 2014/01/16 00:45:12 VCS ERROR V-16-10001-1003 (d-node9) DiskGroup:md_DG:online:** ERROR: vxdg import (force) failed on Disk Group md_DG 2014/01/16 00:45:12 VCS ERROR V-16-10001-1004 (d-node9) DiskGroup:md_DG:online:** ERROR: vxdg import failed on Disk Group md_DG after vxdctl enable 2014/01/16 00:45:13 VCS INFO V-16-2-13716 (d-node9) Resource(mdm_DG): Output of the completed operation (online) ============================================== VxVM vxdg ERROR V-5-1-10978 Disk group md_DG: import failed: No valid disk found containing disk group VxVM vxdg ERROR V-5-1-10978 Disk group md_DG: import failed: No valid disk found containing disk group VxVM vxdg ERROR V-5-1-10978 Disk group md_DG: import failed: No valid disk found containing disk group ============================================== Looking into the old dicussions I have tried this but still didnt work...: changed before importing # vxdmpadm settune dmp_cache_open=off Tunable value will be changed immediately # vxdmpadm gettune all | grep cache dmp_cache_open off on But I got the same error. Same thing o var/adm/mesages: Jan 16 00:45:12 d-node9 Had[17566]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1003 (dp-node9) DiskGroup:mdm_DG:online:** ERROR: vxdg import (force) failed on Disk Group mdm_DG Jan 16 00:45:12 d-node9 vxdmp: [ID 631182 kern.notice] NOTICE: VxVM vxdmp V-5-0-0 removed disk array DISKS, datype = Disk Jan 16 00:45:12 d-node9 vxdmp: [ID 803759 kern.notice] NOTICE: VxVM vxdmp V-5-0-34 added disk array DISKS, datype = Disk Jan 16 00:45:12 d-node9 Had[17566]: [ID 702911 daemon.notice] VCS ERROR V-16-1-1004 (dp-node9) DiskGroup:mdm_DG:online:** ERROR: vxdg import failed on Disk Group mdm_DG after vxdctl enable Any suggestion? Tks, JoaoSolved2.2KViews3likes9CommentsVCS Cluster not starting.
Hello All, I am having difficulties trying to get VCS started on this system. I have attached what I have got so far. I apperciate any comments or suggestions as to go from here. Thank you The hostnames in the main.cf corrosponds to that of the servers. hastatus -sum VCS ERROR V-16-1-10600 Cannot connect to VCS engine VCS WARNING V-16-1-11046 Local system not available hasys -state VCS ERROR V-16-1-10600 Cannot connect to VCS engine hastop -all -force VCS ERROR V-16-1-10600 Cannot connect to VCS engine hastart / hastart -onenode dmesg: Exiting: Another copy of VCS may be running engine_A.log 2013/10/22 15:16:43 VCS NOTICE V-16-1-11051 VCS engine join version=4.1000 2013/10/22 15:16:43 VCS NOTICE V-16-1-11052 VCS engine pstamp=4.1 03/03/05-14:58:00 2013/10/22 15:16:43 VCS NOTICE V-16-1-10114 Opening GAB library 2013/10/22 15:16:43 VCS NOTICE V-16-1-10619 'HAD' starting on: db1 2013/10/22 15:16:45 VCS INFO V-16-1-10125 GAB timeout set to 15000 ms 2013/10/22 15:17:00 VCS CRITICAL V-16-1-11306 Did not receive cluster membership, manual intervention may be needed for seeding #gabconfig -a GAB Port Memberships =============================================================== #lltstat -nvv LLT node information: Node State Link Status Address * 0 db1 OPEN bge1 UP 00:03:BA:15 bge2 UP 00:03:BA:15 1 db2 CONNWAIT bge1 DOWN bge2 DOWN bash-2.05$ lltconfig LLT is running ps -ef | grep had root 826 1 0 15:16:43 ? 0:00 /opt/VRTSvcs/bin/had root 836 1 0 15:16:45 ? 0:00 /opt/VRTSvcs/bin/hashadowSolved18KViews3likes4CommentsHow to manually edit the main.cf file
Hi, I need to make a veryminor edit to the main.cf file. I am assuming this is the way to do this: haconf -makerw edit main.cf on one of the nodes i.e vi /etc/VRTSvcs/conf/config/main.cf save changes hacf -verify /etc/VRTSvcs/conf/config haconf -dump -makeroSolved7.1KViews3likes6CommentsVCS engine (had) cann't start automatically
this is the only node of the cluster1 thatI amtrying to start VCS. as my experience, after SFHA installation, the veritas can start automatically with Solaris OS boot. the below is the procedure of my operation 1.after SFHA installation, I found it cann't run "vxdisk list" command, it showing "Configuration daemon is not accessible" so I follow the article and solved it https://www.veritas.com/support/en_US/article.TECH67777 2.then, i completed the synchronization between cluster1 and cluster2 3.when Itry to config GCO with gcoconfig, it cann't work as well.at that time,I reboot the Solaris OS once and found it persist after reboot. so I try to start veritas with command "hastart -onenode" then I can config the GCO. 4.I completed all veritas configuration cluster1 and cluster 2 both. 5.I reboot the node of cluster1 and cluster2, after reboot, I found only the host of cluster1 cannot accese by VCS_Cluster_Manager_Java_Console. it show VCS ERROR V-16-10-106. But the host of cluster2 is fine. pleaseadvice me.thanks in advance. the detail information has been attached in files.Solved1.6KViews2likes4CommentsRegarding resource online operations
Hi, Suppose, a resouce got faulted in a SG and I need to make it online. Shall I have to do in this way? hagrp -clear <group> [-sys] <host> <sys> and then online the service group hagrp -online <group> -sys <sys> OR by this way? 1.Flush the SG, hagrp -flush <group> -sys <system> 2.Clear the faulted resource hares -clear <resource> [-sys] 3.Online it, hares -online <resource> [-sys] Do flushing of the SG required? Kindly assist how should i have to proceed for these cases.Solved4.7KViews2likes3Comments