Migrate ApplicationHA to new vCenter server
We're consolidating a number of vCenter servers into a single instance. One of the clusters we want to move has VMs managed by AppHA 6.0.1. Is there an easy way to migrate the cluster configuration from the old vCenter server to the new one without causing any cluster outages or losing the cluster configuration? I've found this article on how to deal with the permanent loss of a vCenter server, but an unsure what might happen if the existing vCenter server stays online. https://sort.symantec.com/public/documents/appha/6.0/windows/productguides/html/appha_userguide_60_win/apas11.htmSolved8.6KViews0likes9CommentsCannot unload GAB and LLT on RHEL 6.0
Hi all, i have next problem # lltstat -n LLT node information: Node State Links 0 srv-n1 OPEN 2 * 1 srv-n2 OPEN 2 # gabconfig -a GAB Port Memberships =============================================================== Port a gen 7b4d01 membership 01 Port b gen 7b4d05 membership 01 Port d gen 7b4d04 membership 01 Port h gen 7b4d11 membership 01 # /opt/VRTSvcs/bin/haconf -dump -makero VCS WARNING V-16-1-10369 Cluster not writable. # /opt/VRTSvcs/bin/hastop -all -force # /etc/init.d/vxfen stop Stopping vxfen.. Stopping vxfen.. Done # /etc/init.d/gab stop Stopping GAB: ERROR! Cannot unload GAB module. Clients still exist Kill/Stop clients corresponding to following ports. GAB Port Memberships =============================================================== Port d gen 7b4d04 membership 01 # /etc/init.d/llt stop Stopping LLT: LLT lltconfig ERROR V-14-2-15121 LLT unconfigure aborted, unregister 3 port(s) LLT:Warning: lltconfig failed. Retrying [1] LLT lltconfig ERROR V-14-2-15121 LLT unconfigure aborted, unregister 3 port(s) LLT:Warning: lltconfig failed. Retrying [2] LLT lltconfig ERROR V-14-2-15121 LLT unconfigure aborted, unregister 3 port(s) LLT:Warning: lltconfig failed. Retrying [3] LLT lltconfig ERROR V-14-2-15121 LLT unconfigure aborted, unregister 3 port(s) LLT:Warning: lltconfig failed. Retrying [4] LLT lltconfig ERROR V-14-2-15121 LLT unconfigure aborted, unregister 3 port(s) LLT:Warning: lltconfig failed. Retrying [5] LLT:Error: lltconfig failed OK, i see lsmod and modinfo for details... # lsmod Module Size Used by vxodm 206291 1 vxgms 284352 0 vxglm 289848 0 gab 283317 4 llt 180985 5 gab autofs4 27683 3 sunrpc 241630 1 dmpCLARiiON 11771 1 dmpap 9390 1 vxspec 3174 6 vxio 3261814 1 vxspec vxdmp 377776 20 vxspec,vxio cpufreq_ondemand 10382 1 acpi_cpufreq 8593 3 freq_table 4847 2 cpufreq_ondemand,acpi_cpufreq ipv6 321209 60 vxportal 5940 0 fdd 53457 1 vxodm vxfs 2957815 2 vxportal,fdd exportfs 4202 0 serio_raw 4816 0 i2c_i801 11190 0 iTCO_wdt 11708 0 iTCO_vendor_support 3022 1 iTCO_wdt ioatdma 57872 9 dca 7099 1 ioatdma i5k_amb 5039 0 hwmon 2464 1 i5k_amb i5000_edac 8833 0 edac_core 46055 3 i5000_edac sg 30186 0 shpchp 33448 0 e1000e 140051 0 ext4 353979 3 mbcache 7918 1 ext4 jbd2 89033 1 ext4 dm_mirror 14003 1 dm_region_hash 12200 1 dm_mirror dm_log 10088 3 dm_mirror,dm_region_hash sr_mod 16162 0 cdrom 39769 1 sr_mod sd_mod 37221 18 crc_t10dif 1507 1 sd_mod pata_acpi 3667 0 ata_generic 3611 0 ata_piix 22588 0 ahci 39105 4 qla2xxx 280129 24 scsi_transport_fc 50893 1 qla2xxx scsi_tgt 12107 1 scsi_transport_fc radeon 797054 1 ttm 46942 1 radeon drm_kms_helper 32113 1 radeon drm 200778 3 radeon,ttm,drm_kms_helper i2c_algo_bit 5664 1 radeon i2c_core 31274 5 i2c_i801,radeon,drm_kms_helper,drm,i2c_algo_bit dm_mod 76856 20 dm_mirror,dm_log # rmmod gab ERROR: Module gab is in use [root@srv-vrts-n2 ~]# modinfo gab filename: /lib/modules/2.6.32-71.el6.x86_64/veritas/vcs/gab.ko license: Proprietary. Send bug reports to support@veritas.com description: Group Membership and Atomic Broadcast 5.1.120.000-SP1PR2 author: VERITAS Software Corp. srcversion: F43C75576C05662FB0ED8C8 depends: llt vermagic: 2.6.32-71.el6.x86_64 SMP mod_unload modversions parm: gab_logflag:int parm: gab_numnids:maximum nodes in the cluster (1-128) (int) parm: gab_numports:maximum gab ports allowed (1-32) (int) parm: gab_flowctrl:queue depth that causes flow-control (1-128) (int) parm: gab_logbufsize:internal log buffer size in bytes (8100-65400) (int) parm: gab_msglogsize:maximum messages in internal message log (128-4096) (int) parm: gab_isolate_time:maximum time to wait for isolated client (16000-240000) (int) parm: gab_kill_ntries:number of times to attempt to kill client (3-10) (int) parm: gab_kstat_size:Number of system statistics to maintain in GAB 60-240 (int) parm: gab_conn_wait:maximum number of wait for CONNECTS message (1-256) (int) parm: gab_ibuf_count:maximum number of intermediate buffers (0-32) (int) # modinfo llt filename: /lib/modules/2.6.32-71.el6.x86_64/veritas/vcs/llt.ko license: Proprietary. Send bug reports to support@veritas.com author: VERITAS Software Corp. description: Low Latency Transport 5.1.120.000-SP1PR2 srcversion: AF11D9C04A71073E1ADCFC8 depends: vermagic: 2.6.32-71.el6.x86_64 SMP mod_unload modversions parm: llt_maxnids:maximum nodes in the cluster (1-128) (int) parm: llt_maxports:maximum llt ports allowed (1-32) (int) parm: llt_nqthread:number of kernel threads to use (2-5) (int) parm: llt_basetimer:frequency of base timer ((10 * 1000)-(500 * 1000)) (int) Hm.... ok, i run /etc/init.d/gab stop with debug ... + echo 'Stopping GAB: ' Stopping GAB: + mod_isloaded ++ lsmod ++ grep '^gab\ ' + return + mod_isconfigured ++ LANG=C ++ LC_ALL=C ++ /sbin/gabconfig -l ++ grep 'Driver state' ++ grep -q Configured + return + /sbin/gabconfig -U + ret=1 + '[' '!' 1 -eq 0 ']' + echo 'ERROR! Cannot unload GAB module. Clients still exist' ERROR! Cannot unload GAB module. Clients still exist + echo 'Kill/Stop clients corresponding to following ports.' Kill/Stop clients corresponding to following ports. + LANG=C + LC_ALL=C + /sbin/gabconfig -a + grep -v 'Port a gen' GAB Port Memberships ... ok, i use /sbin/gabconfig -l and -U # /sbin/gabconfig -l GAB Driver Configuration Driver state : Configured Partition arbitration: Disabled Control port seed : Enabled Halt on process death: Disabled Missed heartbeat halt: Disabled Halt on rejoin : Disabled Keep on killing : Disabled Quorum flag : Disabled Restart : Enabled Node count : 2 Send queue limit : 128 Recv queue limit : 128 IOFENCE timeout (ms) : 15000 Stable timeout (ms) : 5000 # /sbin/gabconfig -U GAB /sbin/gabconfig ERROR V-15-2-25014 clients still registered but it did not help... i find this topic https://www-secure.symantec.com/connect/forums/unable-stop-gab-llt-vcs51solaris-10 and stop ODM # /etc/init.d/vxodm stop Stopping ODM and run /etc/init.d/gab stop , but # /etc/init.d/gab stop Stopping GAB: GAB has usage count greater than zero. Cannot unload I again see /sbin/gabconfig -l # /sbin/gabconfig -l GAB Driver Configuration Driver state : Unconfigured Partition arbitration: Disabled Control port seed : Disabled Halt on process death: Disabled Missed heartbeat halt: Disabled Halt on rejoin : Disabled Keep on killing : Disabled Quorum flag : Disabled Restart : Disabled Node count : 0 Send queue limit : 128 Recv queue limit : 128 IOFENCE timeout (ms) : 15000 Stable timeout (ms) : 5000 [root@srv-vrts-n2 ~]# /sbin/gabconfig -a GAB Port Memberships =============================================================== and again run with debug /etc/init.d/gab stop ... + echo 'Stopping GAB: ' Stopping GAB: + mod_isloaded ++ lsmod ++ grep '^gab\ ' + return + mod_isconfigured ++ LANG=C ++ LC_ALL=C ++ /sbin/gabconfig -l ++ grep 'Driver state' ++ grep -q Configured + return + mod_unload ++ lsmod ++ grep '^gab ' ++ awk '{print $3}' + USECNT=1 + '[' -z 1 ']' + '[' 5 '!=' 0 ']' + ps -e + grep gablogd + '[' 1 -ne 0 ']' + GAB_UNLOAD_RETRIES=0 + '[' 0 '!=' 0 ']' ++ lsmod ++ grep '^gab ' ++ awk '{print $3}' + USECNT=1 + '[' 1 -gt 0 ']' + echo 'GAB has usage count greater than zero. Cannot unload' GAB has usage count greater than zero. Cannot unload + return 1 ... and agian run lsmod lsmod | grep gab gab 283317 1 llt 180985 1 gab [root@srv-vrts-n2 ~]# rmmod gab ERROR: Module gab is in use what can be done in such a situation?Solved8KViews1like4Commentscannot configure vxfen after reboot
Hello, We move physically a server, and after reboot, we cannot configure vxfen. # vxfenconfig -c VXFEN vxfenconfig ERROR V-11-2-1002 Open failed for device: /dev/vxfen with error 2 my vxfen.log : Wed Aug 19 13:17:09 CEST 2015 Invoked vxfen. Starting Wed Aug 19 13:17:23 CEST 2015 return value from above operation is 1 Wed Aug 19 13:17:23 CEST 2015 output was VXFEN vxfenconfig ERROR V-11-2-1041 Snapshot for this node is different from that of the running cluster. Log Buffer: 0xffffffffa0c928a0 VXFEN vxfenconfig NOTICE Driver will use customized fencing - mechanism cps Wed Aug 19 13:17:23 CEST 2015 exiting with 1 Engine version 6.0.10.0 RHEL 6.3 any idea to help me running the vxfen (and the had after ... ) ?6.2KViews0likes7CommentsRequirements for my Master Server (RAM memory and processors)
Hi all. I have great doubt I am implementing a new platform in which there are 15 clients (Sun X4270). the questions is: The master server with netbackup to manage clients, that memory (RAM) should have? the ram of the master server depends on the clients? And talking about processors, also depends on customer? Obviously already taking into account the necessities of the installed applications Thank you and good daySolved6KViews1like9Commentsvxdisk list showing errors on multiple disks, and I am unable to start cluster on slave node.
Hello, If anybody have same experience and can help me, I am gonna be very thankful I am using solars 10 (x86 141445-09) + EMC PowerPath (5.5.P01_b002) + vxvm (5.0,REV=04.15.2007.12.15) on two node cluster. This is fileserver cluster. I've added couple new LUNs and when I try to scan for new disk :"vxdisk scandisks" command hangs and after that time I was unable to do any vxvm job on that node, everytime command hangs. I've rebooted server in maintanance windows, (before reboot switched all SGs on 2nd node) After that reboot I am unable to join to cluster with reason 2014/04/13 01:04:48 VCS WARNING V-16-10001-1002 (filesvr1) CVMCluster:cvm_clus:online:CVMCluster start failed on this node. 2014/04/13 01:04:49 VCS INFO V-16-2-13001 (filesvr1) Resource(cvm_clus): Output of the completed operation (online) ERROR: 2014/04/13 01:04:49 VCS ERROR V-16-10001-1005 (filesvr1) CVMCluster:???:monitor:node - state: out of cluster reason: Cannot find disk on slave node: retry to add a node failed Apr 13 01:10:09 s_local@filesvr1 vxvm: vxconfigd: [ID 702911 daemon.warning] V-5-1-8222 slave: missing disk 1306358680.76.filesvr1 Apr 13 01:10:09 s_local@filesvr1 vxvm: vxconfigd: [ID 702911 daemon.warning] V-5-1-7830 cannot find disk 1306358680.76.filesvr1 Apr 13 01:10:09 s_local@filesvr1 vxvm: vxconfigd: [ID 702911 daemon.error] V-5-1-11092 cleanup_client: (Cannot find disk on slave node) 222 here is output from 2nd node (working fine) Disk: emcpower33s2 type: auto flags: online ready private autoconfig shared autoimport imported guid: {665c6838-1dd2-11b2-b1c1-00238b8a7c90} udid: DGC%5FVRAID%5FCKM00111001420%5F6006016066902C00915931414A86E011 site: - diskid: 1306358680.76.filesvr1 dgname: fileimgdg dgid: 1254302839.50.filesvr1 clusterid: filesvrvcs info: format=cdsdisk,privoffset=256,pubslice=2,privslice=2 and here is from node where i see this problems Device: emcpower33s2 devicetag: emcpower33 type: auto flags: error private autoconfig pubpaths: block=/dev/vx/dmp/emcpower33s2 char=/dev/vx/rdmp/emcpower33s2 guid: {665c6838-1dd2-11b2-b1c1-00238b8a7c90} udid: DGC%5FVRAID%5FCKM00111001420%5F6006016066902C00915931414A86E011 site: - errno: Configuration request too large Multipathing information: numpaths: 1 emcpower33c state=enabled Can anybody help me? I am not sure about Configuration request too largeSolved5.9KViews1like16CommentsIP agent for same mac address interface
Hi all, Our environment as the following: OS: redhat 6.5 VCS: VCS 6.2.1 Our server have two physical network port, namely eth0 and eth1. We do create tagged vlan, vlan515, vlan516, vlan518, vlan520 based on eth0 and eth1. We are able to create resource IP on vlan518 and failover between two nodes. However, when we create resource IP on vlan515, it is not able to bring it online. According to the link, https://support.symantec.com/en_US/article.TECH214469.html, It knows that duplicate mac address would cause the problem. However, it can't figure out where "MACAddress" attribute in VCS Java Console as mentioned in the solution. I did manually add "MACAddress" attribute on main.cf on either NIC or IP resource, it come with not support with haconf -verify command. Any hints or solution for the problem when configure the IP agent resource on same mac address? Thanks, XentarSolved5.6KViews0likes22CommentsCreate second RDS using volumes from same disk group in Veritas volume replicator
Hi all, I have one query. I have one disk group with 5 volumes, which are more than 2 TB and are configured for replication. Now i want to add 2 more volumes in the RDS but it will resync all the volumes after adding these 2 volumes, so is it possible to make second RDS for only those 2 new volumes. Also can we use same IP for replication.5.5KViews0likes8CommentsHow to recover from failed attempt to switch to a different node in cluster
Hello everyone. I have a two node cluster to serve as an Oracle database server. I have the Oracle binaries installed on disks local to each of the nodes (so they are outside the control of the Cluster Manager). I have a diskgroup which is three 1TB LUNs from my SAN, six volumes on the diskgroup (u02 through u07), six mount points (/u02 through /u07), a database listener and the actual Oracle database. I was able to successfully manually bring up these individual components and confirmed that the database was up an running. I then tried a "Switch To" operation to see if everything would mode to the other node of the cluster. It turns out this was a bad idea. Within the Cluster Manager gui, the diskgroup has a state of Online, Istate of "Waiting to go offline propogate" and Flag of "Unable to offline". The volumes show as "Offline on all systems" but the mounts still show as online with "Status Unknown". When I try to take the mount points offline, I get the message "VCS ERROR V-16-1-10277 The Service Group i1025prd to which Resource Mnt_scratch belongs has failed or switch, online, and offline operations are prohibited." Can anyone tell me how I can fix this? KenSolved5KViews1like17CommentsSF4.1 VxDMP disables dmpnode on single path failure
This is more like an informational question, since I do not assume anyone has a solution, but just in case I would be thankfull for some enlightment: I am forced to use an old Version SF4.1MP4 in this case on Linux SLES9. For whatever reason DMP does not work with the JBOD I have added. The JBOD (Promise VTrak 610fD) is ALUA, so half of all the available paths are always standy and that is ok. But the DMP in 4.1 when seeing one of 4 paths not working diables the whole DMP Node, rendering the disk unusable: Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-148 i/o error occured on path 8/0x70 belonging to dmpnode 201/0x10<5>VxVM vxdmp V-5-0-148 i/o error anal ysis done on path 8/0x70 belonging to dmpnode 201/0x10<5>VxVM vxdmp V-5-0-0 SCSI error opcode=0x28 returned status=0x1 key=0x2 asc=0x4 ascq=0xb on path 8/0x70 Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-112 disabled path 8/0x50 belonging to the dmpnode 201/0x10 Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-112 disabled path 8/0x70 belonging to the dmpnode 201/0x10 Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-112 disabled path 8/0x10 belonging to the dmpnode 201/0x10 Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-112 disabled path 8/0x30 belonging to the dmpnode 201/0x10 Jun 12 15:35:01 kernel: VxVM vxdmp V-5-0-111 disabled dmpnode 201/0x10 Jun 12 15:35:01 kernel: Buffer I/O error on device VxDMP2, logical block 0 Currently my only solutions seems to stick with Linux DM-Multipathing an add the disks as foreign devices.SolvedDMP, MPIO, MSDSM, SCSI-3 and ALUA configuration settings
Ok, I'm somewhat confused and the more I read the more confused I think I'm getting. I'm going to be setting up a 4 node active/active cluster for SQL. All of the nodes will have 2 seperate fiber channel HBAs connecting through 2 seperate switches to our NetApp. The NetApp supports ALUA, so the storage guy wants to use it. It is my understanding that I need to use SCSI-3 to get this to work. Sounds good to me so far. My question is, do I need to use any of Microsoft's MPIO or MSDSM? This is on Win 2008 R2. Or does Veritas take care of all of that? Also, I read that in a new cluster set up, only connect 1 path first and then install and then connect the 2nd path and let Veritas detect it and configure it. Is that accurate? Any info or directions you can point me will be greatly appreciated. Thanks!Solved