VCS cannot startup
Hi All , The enviornment is configured two node form the active / passive cluster, i have maintenance for active node , switch to passive node to online cluster, but check the status is in parital, I have no idea what happen on the issue. Can you please advice how to fix it ? root@devuaebms42 # gabconfig -a GAB Port Memberships =============================================================== Port a gen 63750d membership 01 Port h gen 63750b membership ;1 Port h gen 63750b visible 0 ^Croot@devuaebms42 # hastatus -sum -- SYSTEM STATE -- System State Frozen A devuaebms41 EXITED 0 A devuaebms42 RUNNING 0 -- GROUP STATE -- Group System Probed AutoDisabled State B cf_bms_sg_01 devuaebms41 Y Y OFFLINE B cf_bms_sg_01 devuaebms42 Y N PARTIAL Many thanks, HongSolved2.8KViews0likes9CommentsVCS Cluster not starting.
Hello All, I am having difficulties trying to get VCS started on this system. I have attached what I have got so far. I apperciate any comments or suggestions as to go from here. Thank you The hostnames in the main.cf corrosponds to that of the servers. hastatus -sum VCS ERROR V-16-1-10600 Cannot connect to VCS engine VCS WARNING V-16-1-11046 Local system not available hasys -state VCS ERROR V-16-1-10600 Cannot connect to VCS engine hastop -all -force VCS ERROR V-16-1-10600 Cannot connect to VCS engine hastart / hastart -onenode dmesg: Exiting: Another copy of VCS may be running engine_A.log 2013/10/22 15:16:43 VCS NOTICE V-16-1-11051 VCS engine join version=4.1000 2013/10/22 15:16:43 VCS NOTICE V-16-1-11052 VCS engine pstamp=4.1 03/03/05-14:58:00 2013/10/22 15:16:43 VCS NOTICE V-16-1-10114 Opening GAB library 2013/10/22 15:16:43 VCS NOTICE V-16-1-10619 'HAD' starting on: db1 2013/10/22 15:16:45 VCS INFO V-16-1-10125 GAB timeout set to 15000 ms 2013/10/22 15:17:00 VCS CRITICAL V-16-1-11306 Did not receive cluster membership, manual intervention may be needed for seeding #gabconfig -a GAB Port Memberships =============================================================== #lltstat -nvv LLT node information: Node State Link Status Address * 0 db1 OPEN bge1 UP 00:03:BA:15 bge2 UP 00:03:BA:15 1 db2 CONNWAIT bge1 DOWN bge2 DOWN bash-2.05$ lltconfig LLT is running ps -ef | grep had root 826 1 0 15:16:43 ? 0:00 /opt/VRTSvcs/bin/had root 836 1 0 15:16:45 ? 0:00 /opt/VRTSvcs/bin/hashadowSolved18KViews3likes4CommentsDoubts on VxVM,VCS Upgrade & root disk encapsulation
Hi All, I have the below queries please 1) In order to stop VxVM to load at system boot time, we need to modify /etc/system file. What entries are to be commented out ? Is it only rootdev:/pseudo/vxio@0:0 set vxio:vol_rootdev_is_volume=1 (OR) also below entries are to be commented out ? forceload: drv/vxdmp forceload: drv/vxio forceload: drv/vxspec 2) My current version of SFHA is 4.1. Once vxfen, gab & llt modules are unloaded to upgrade to 4.1MP2, should i again unload these modules to further upgrade to 5.1SP1 and again to 5.1SP1RP4 (OR) 6.0 ? After each upgrade should i stop the services in /etc/init.d and unload modules (OR) stopping services & unloading modules only once is enough to further upgrade to other versions ? My Plan is to upgrade from 4.1---> 4.1 MP2---> 5.1SP1--->5.1SP1RP4 (OR) 6.0 3) Before upgrading should i also stop & unload the below listed modules 24 12800a8 26920 268 1 vxdmp (VxVM 4.1z: DMP Driver) 25 7be00000 2115c8 269 1 vxio (VxVM 4.1z I/O driver) 27 12a4698 13f0 270 1 vxspec (VxVM 4.1z control/status driver) 213 7b2d7528 c40 272 1 vxportal (VxFS 4.1_REV-4.1B18_sol_GA_s10b) 214 7ae00000 1706a8 20 1 vxfs (VxFS 4.1_REV-4.1B18_sol_GA_s10b) If yes, should i stop & unload after each upgrade (OR) doing it once is enough ? 4) Once the OS comes up with native disks (c#t#d#s#), In order to bring it under VxVM control we need to encapsulate using vxdiskadm. My doubt is will rootdg, rootvol, plexes & subdisks be created automatically? Need a little clarification regarding this please. Response is highly appreciated as always, Thank you very much. Regards, Danish.Solved1.9KViews0likes1CommentVCS Cluster not starting.
Hi I am facing problem while trying to start VCS . From LOG : ============================================================== tail /var/VRTSvcs/log/engine_A.log 2014/01/13 21:39:14 VCS NOTICE V-16-1-11050 VCS engine version=5.1 2014/01/13 21:39:14 VCS NOTICE V-16-1-11051 VCS engine join version=5.1.00.0 2014/01/13 21:39:14 VCS NOTICE V-16-1-11052 VCS engine pstamp=Veritas-5.1-10/06/09-14:37:00 2014/01/13 21:39:14 VCS INFO V-16-1-10196 Cluster logger started 2014/01/13 21:39:14 VCS NOTICE V-16-1-10114 Opening GAB library 2014/01/13 21:39:14 VCS NOTICE V-16-1-10619 ‘HAD’ starting on: nsscls01 2014/01/13 21:39:16 VCS INFO V-16-1-10125 GAB timeout set to 30000 ms 2014/01/13 21:39:16 VCS NOTICE V-16-1-11057 GAB registration monitoring timeout set to 200000 ms 2014/01/13 21:39:16 VCS NOTICE V-16-1-11059 GAB registration monitoring action set to log system message 2014/01/13 21:39:31 VCS CRITICAL V-16-1-11306 Did not receive cluster membership, manual intervention may be needed for seeding ============================================================================================= root@nsscls01# hastatus -sum VCS ERROR V-16-1-10600 Cannot connect to VCS engine VCS WARNING V-16-1-11046 Local system not available Please advice how can I start the VCS.Solved16KViews2likes11Commentsvxconfigd core dumps at vxdisk scandisks after zpool removed from ldom
Hi I'm testing InfoScale 7.0 on Solaris with LDoms. Creating a ZPOOL in the LDom works. It seems there is something not working properly. On the LDom Console I see May 23 16:19:45 g0102 vxdmp: [ID 557473 kern.warning] WARNING: VxVM vxdmp V-5-3-2065 dmp_devno_to_devidstr ldi_get_devid failed for devno 0x11500000000 May 23 16:19:45 g0102 vxdmp: [ID 423856 kern.warning] WARNING: VxVM vxdmp V-5-0-2046 : Failed to get devid for device 0x20928e88 After I destroy the ZPOOL, I would like to remove the Disk from the LDom. To be able to do that I remove and disable the disk /usr/sbin/vxdmpadm -f disable path=c1d1s2 /usr/sbin/vxdisk rm c1d1s2 After this I'm able to remove the Disk from the LDom using ldm remove-vdisk. The dmp configuration is not cleaned up. # /usr/sbin/vxdmpadm getsubpaths ctlr=c1 NAME STATE[A] PATH-TYPE[M] DMPNODENAME ENCLR-TYPE ENCLR-NAME ATTRS ================================================================================ NONAME DISABLED(M) - NONAME OTHER_DISKS other_disks STANDBY c1d0s2 ENABLED(A) - c1d0s2 OTHER_DISKS other_disks - # If I run vxdisk scandisks at this stage, the vxdisk command hangs and the vxconfigd core dumps: # file core core: ELF 32-bit MSB core file SPARC Version 1, from 'vxconfigd' # pstack core core 'core' of 378: vxconfigd -x syslog -m boot ------------ lwp# 1 / thread# 1 --------------- 001dc018 ddl_get_disk_given_path (0, 0, 0, 0, 66e140, 0) 001d4230 ddl_reconfigure_all (49c00, 0, 400790, 3b68e8, 404424, 404420) + 690 001b0bfc ddl_find_devices_in_system (492e4, 3b68e8, 42fbec, 4007b4, 4db34, 0) + 67c 0013ac90 find_devices_in_system (2, 3db000, 3c00, 50000, 0, 3d9400) + 38 000ae630 ddl_scan_devices (3fc688, 654210, 0, 0, 0, 3fc400) + 128 000ae4f4 req_scan_disks (660d68, 44fde8, 0, 654210, ffffffec, 3fc400) + 18 00167958 request_loop (1, 44fde8, 3eb2e8, 1800, 19bc, 1940) + bfc 0012e1e8 main (3d8000, ffbffcd4, ffffffff, 42b610, 0, 33bb7c) + f2c 00059028 _start (0, 0, 0, 0, 0, 0) + 108 Thanks, Marcel1.7KViews0likes1Commentclone disk group
Greetings, I need to migrate disk groups between hosts. The current aging server runs vxvm 5.x on solaris 10. The proposed work loads are to be taken on by a combination of Solaris 11 and Solaris 10 logocal domains, split between application and database. I'm not using vcs only vxvm. Thevxvm version on the new platform is 7.1. Due to limitations on the storage arrays, i cannot create clones on the array and map them to the new host. Does vxvm have a cloning mechanism ? Is there is a better approach to migrate the data across different vxvm versions and maintain a point of failback. I would like to maintain the the diskgroups separately until the cutover. The DG configs: # # app-dg # # Lun Veritas Disk Veritas DiskGroup 6000144000000010A00CB5581BC5169F app-disk0 app-dg 6000144000000010A00CB5581BC51699 app-disk1 app-dg 6000144000000010A00CB5581BC516A6 app-disk2 app-dg # # applocal-dg # # Lun Veritas Disk Veritas DiskGroup 6000144000000010A00CB5581BC5161A applocal-disk0 applocal-dg 6000144000000010A00CB5581BC51626 applocal-disk1 applocal-dg 6000144000000010A00CB5581BC51627 applocal-disk2 applocal-dg 6000144000000010A00CB5581BC51619 applocal-disk3 applocal-dg # # db_ora-dg # # Lun Veritas Disk Veritas DiskGroup 6000144000000010A00CB5581BC5161D db_ora-disk0 db_ora-dg 6000144000000010A00CB5581BC5161C db_ora-disk1 db_ora-dg 6000144000000010A00CB5581BC5161B db_ora-disk2 db_ora-dg 6000144000000010A00CB5581BC515F7 db_ora-disk3 db_ora-dg # # db_ora02-dg # # Lun Veritas Disk Veritas DiskGroup 6000144000000010A00CB5581BC51714 db_ora02-disk1 db_ora02-dg 6000144000000010A00CB5581BC51719 db_ora02-disk2 db_ora02-dg 6000144000000010A00CB5581BC5170D db_ora02-disk3 db_ora02-dg cheers MB2.4KViews0likes3CommentsSolaris 11.1 VXVM 6.0.1 'df' cause a panic
Environment: System Configuration: HP ProLiant BL480c G1 Oracle Solaris 11.1 X86 panic string: BAD TRAP: type=e (#pf Page fault) rp=fffffffc816fdb90 addr=0 occurred in module "unix" due to a NULL pointer dereference Veritas INFO: PKGINST: VRTSvxvm NAME: Binaries for VERITAS Volume Manager by Symantec CATEGORY: system ARCH: i386 VERSION: 6.0.100.000,REV=08.01.2012.08.52 Stack: genunix: [ID 655072 kern.notice] fffffffc816fdab0 unix:die+105 () genunix: [ID 655072 kern.notice] fffffffc816fdb80 unix:trap+153e () genunix: [ID 655072 kern.notice] fffffffc816fdb90 unix:cmntrap+e6 () genunix: [ID 655072 kern.notice] fffffffc816fdca0 unix:strncpy+1c () genunix: [ID 655072 kern.notice] fffffffc816fdcd0 odm:odmstatvfs+90 () genunix: [ID 655072 kern.notice] fffffffc816fdcf0 genunix:fsop_statfs+1a () genunix: [ID 655072 kern.notice] fffffffc816fde70 genunix:cstatvfs64_32+42 () genunix: [ID 655072 kern.notice] fffffffc816fdec0 genunix:statvfs64_32+69 () genunix: [ID 655072 kern.notice] fffffffc816fdf10 unix:brand_sys_sysenter+1dc () Messages: unix: [ID 839527 kern.notice] df: unix: [ID 753105 kern.notice] #pf Page fault unix: [ID 532287 kern.notice] Bad kernel fault at addr=0x0 unix: [ID 243837 kern.notice] pid=3965, pc=0xfffffffffb893ff8, sp=0xfffffffc816fdc88, eflags=0x10206 unix: [ID 211416 kern.notice] cr0: 80050033<pg,wp,ne,et,mp,pe> cr4: 6f8<xmme,fxsr,pge,mce,pae,pse,de> unix: [ID 624947 kern.notice] cr2: 0 unix: [ID 625075 kern.notice] cr3: 59f0a2000 unix: [ID 625715 kern.notice] cr8: c unix: [ID 100000 kern.notice] unix: [ID 592667 kern.notice] rdi: fffffffc816fdd48 rsi: 0 rdx: f unix: [ID 592667 kern.notice] rcx: 1 r8: e80 r9: 0 unix: [ID 592667 kern.notice] rax: fffffffc816fdd48 rbx: fefa3430 rbp: fffffffc816fdca0 unix: [ID 592667 kern.notice] r10: fffffffffb856d00 r11: 0 r12: fffffffc816fdd00 unix: [ID 592667 kern.notice] r13: ffffc10012176880 r14: 0 r15: ffffc1002bb09480 unix: [ID 592667 kern.notice] fsb: 0 gsb: ffffc1000eac8000 ds: 4b unix: [ID 592667 kern.notice] es: 4b fs: 0 gs: 1c3 unix: [ID 592667 kern.notice] trp: e err: 0 rip: fffffffffb893ff8 unix: [ID 592667 kern.notice] cs: 30 rfl: 10206 rsp: fffffffc816fdc88 unix: [ID 266532 kern.notice] ss: 38 In preced log of panic I see "odm:odmstatvfs+90". I think this is root of panic, but due in lack of scat and mdb knowlage, I am cannot to investigate this module. When I delete VXVM, there is no panic when I issue the 'df'. If I can provide more information about this case, please let me know. For now I dont know what additional info to provide. Core dump is about of 400 MB, which is more than I can attach to this message.Solved4KViews1like13Commentsunsuccessful cluster failover occured because of nic faulted
Hello, Platform: solaris 11 Logs: Jun 23 16:36:49 nodeA Had[5211]: [ID 702911 daemon.notice] VCS ERROR V-16-1-54031 Resource csgnic (Owner: Unspecified, Group: ClusterService) is FAULTED on sys nodeA Jun 23 16:37:49 nodeA Had[5211]: [ID 702911 daemon.notice] VCS ERROR V-16-1-54031 Resource nic_proxy_aggr1 (Owner: Unspecified, Group: oracle) is FAULTED on sys nodeA Question 1) I need more detail about the problem. I tried to check /var/log/messages, /var/adm/messages, /var/fm/fmd/*files and I can’t see anything related with this error. Which logs should be checked on the solaris 11 system for this situation? Question 2) What kind of method do you advise to investigate the nic problem forgetting more information on platform? Question 3) What kind of configuration should I do for handling with nic failures?Solved3.3KViews1like7Commentsupgrade SFHA
Hi All, Regarding the need of configuring DR with VVR we have purchase the new licence Infoscale 7.1 Can we upgrade from SFHA 6.1 to Veritas Infoscale Enterprise 7.1? if not what are the best method to preserve data when installing the new product. Thanks in advance.1.8KViews0likes0Comments