Automatic Failover
I would like to use Veritas Cluster Server to achieve high availability and automatic fail over for my applications. These application are java applications with some services with the backend databases using Sybase The java applications can be broken into 2 parts : web layer and application layer. The underlying infrastructure will be in RHEL Linux for java applications and database (Sybase) My question is : is VCS supporting seamless automatic failover for the java services including database services without requiring manual intervention ? What i want to achieve is : after i setup active-passive for application layer, i expect the active node to automatically failover to passive node and immediately the passive node become active nodeSolved998Views0likes1Commentdeleting rlink that's having "secondary_config_err" flag
hello, in my VCS global cluster my ORAGrp resource group is partially online since my rvgres is offline, i am suspecting the issue in the below rlink. i am trying to dissociate thi rlink(below:rlk_sys1-DB-rep_DB_r) and dettach it in order to delete it but i am not able to succeed. below are some output from the system. root@sys2# vxprint -P Disk group: DBhrDG TY NAME ASSOC KSTATE LENGTH PLOFFS STATE TUTIL0 PUTIL0 rl rlk_sys1-DB-rep_DB_r DB_rvg CONNECT - - ACTIVE - - rl rlk_sys1-rep_DB-rvg DB-rvg ENABLED - - PAUSE - - root@sys2# vxrlink -g DBhrDG dis rlk_sys1-rep_DB-rvg VxVM VVR vxrlink ERROR V-5-1-3520 Rlink rlk_sys1-rep_DB-rvg can not be dissociated if it is attached root@sys2# vxrlink -g DBhrDG det rlk_sys1-rep_DB-rvg VxVM VVR vxrlink ERROR V-5-1-10128 Operation not allowed with attached rlinks root@sys2# vxedit -g DBhrDG rm rlk_sys1-rep_DB-rvg VxVM vxedit ERROR V-5-1-3540 Rlink rlk_sys1-rep_DB-rvg is not disabled, use -f flag root@sys2# vxedit -g DBhrDG -f rm rlk_sys1-rep_DB-rvg VxVM vxedit ERROR V-5-1-3541 Rlink rlk_sys1-rep_DB-rvg is not dissociated root@sys2# vxprint -Vl Disk group: DBhrDG Rvg: DB-rvg info: rid=0.1317 version=0 rvg_version=41 last_tag=11 state: state=CLEAN kernel=DISABLED assoc: datavols=(none) srl=(none) rlinks=rlk_sys1-rep_DB-rvg exports=(none) vsets=(none) att: rlinks=rlk_sys1-rep_DB-rvg flags: closed secondary disabled detached passthru logging device: minor=26012 bdev=343/26012 cdev=343/26012 path=/dev/vx/dsk/DBhrDG/DB-rvg perms: user=root group=root mode=0600 Rvg: DB_rvg info: rid=0.1386 version=13 rvg_version=41 last_tag=12 state: state=ACTIVE kernel=ENABLED assoc: datavols=sys1_DB_Process,sys1_DB_Script,... srl=sys1_DB_SRL rlinks=rlk_sys1-DB-rep_DB_r exports=(none) vsets=(none) att: rlinks=rlk_sys1-DB-rep_DB_r flags: closed secondary enabled attached logging device: minor=26014 bdev=343/26014 cdev=343/26014 path=/dev/vx/dsk/DBhrDG/DB_rvg perms: user=root group=root mode=0600 please advise regards1.4KViews0likes4CommentsStorage Foundation
Hi im new to the product and the field, so really any advice would be great. where exactly do you install VOM in your SAN environment and does the SF s/w collide with the storage vendor management s/w. And how do you control a heterogenous storage with SF Also can someone explain the licensing.Solvedadding new volumes to a DG that has a RVG under VCS cluster
hi, i am having a VCS cluster with GCO and VVR. on each node of the cluster i have a DG with an associated RVG, this RVG contains 11 data volume for Oracle database, these volumes are getting full so i am going to add new disks to the DG and create new volumes and mount points to be used by the Oracle Database. my question:can i add the disks to the DG and volumes to RVGwhile the database is UP and the replication is ON? if the answer is no, please let me know what should be performed on the RVG and rlinkto add these volumes also what to perform on the database resource group to not failover. thanks in advance.Solved4.4KViews0likes14CommentsTrigger after failed cleanup script
Hi there, I have a system where the cleanup script can fail/timeout and I want to execute another script if this happens. And I was wondering which can be the best way of doing this. In the veritas cluster server administrators guide for Linux I found the trigger RESNOTOFF. From the documentation it is my understanding that this trigger will be triggered in the following cases: A resource fails going offline (started by VCS) and the clean up fails. A resource goes offline unexpectedly and the clean up fails. I have tested this and the RESNOTOFF is working in the first scenario but not in the second. For testing the second scenario I kill the service and I can see the following message in the engine_A.log: VCS ERROR V-16-2-13067 (node1) Agent is calling clean for resource(service1) because the resource became OFFLINE unexpectedly, on its own. When the cleanup fails I would expect the resource to became UNABLE TO OFFLINE. However, the status of the resource is still ONLINE: # hares -state service1 #Resource Attribute System Value service1 State node1 ONLINE service1 State node2 OFFLINE So the resource is ONLINE and VCS keeps running the cleanup command indefinitely (which is failing). I was wondering if I need to configure something else to make the RESNOTOFF to work in this particular scenario. Thanks,Solved938Views0likes3CommentsVXVM gives trouble in solaris11.2
Hi, I have installed VXVM6.0.1SF with trial version on Solaris11.2 inVMWare workstation. While installing the installer asked to set lwp_default and stksize to 0x8000. I have done it and completed installation successfully without any issue. After reboot also no issue. But if i give df -h command then immediately system going to kernel panic and starts rebooting and while booting it throws warning in lwp_default and stksize. Also i can see logs to run the vxkeyless setnone. This also done. Need any work around to fix this issue. Regards, vinothtvSolved1.3KViews1like1CommentReplacing/Restoring Failed Drive
Yestarday, I had two drives in my storage san fail. The san monitor reports that all the physical drives are fine. bash-2.03# vxdisk list | grep AMS AMS_WMS0_0 auto:cdsdisk remote996006 remote9960 online AMS_WMS0_1 auto:cdsdisk remote996004 remote9960 online AMS_WMS0_2 auto:cdsdisk remote996013 remote9960 online AMS_WMS01_0 auto - - error AMS_WMS01_1 auto - - error AMS_WMS01_2 auto:cdsdisk remote996012 remote9960 online AMS_WMS012_0 auto:cdsdisk remote996008 remote9960 online AMS_WMS012_1 auto:cdsdisk remote996002 remote9960 online AMS_WMS012_2 auto:cdsdisk remote996011 remote9960 online AMS_WMS0123_0 auto:cdsdisk remote996007 remote9960 online AMS_WMS0123_1 auto:cdsdisk remote996001 remote9960 online AMS_WMS0123_2 auto:cdsdisk remote996010 remote9960 online - - remote996003 remote9960 failed was:AMS_WMS01_1 - - remote996005 remote9960 failed was:AMS_WMS01_0 The devices on lines 5/6 are new. A result of me attempting to repair the two failed drives at the bottom. So far i've unmounted the volumes associated with the drives, but there are 11 more mounts attached to the Disk Group. Being new, I'm not sure what processes and systems are using those volumes and am reluctant to unmount them at the moment. I followed several guides trying to determine the cause of the problem and/or to restore the two disks. Here's some of the results of my effort so far. > vxprint -htg remote9960 DG NAME NCONFIG NLOG MINORS GROUP-ID ST NAME STATE DM_CNT SPARE_CNT APPVOL_CNT DM NAME DEVICE TYPE PRIVLEN PUBLEN STATE RV NAME RLINK_CNT KSTATE STATE PRIMARY DATAVOLS SRL RL NAME RVG KSTATE STATE REM_HOST REM_DG REM_RLNK CO NAME CACHEVOL KSTATE STATE VT NAME NVOLUME KSTATE STATE V NAME RVG/VSET/CO KSTATE STATE LENGTH READPOL PREFPLEX UTYPE PL NAME VOLUME KSTATE STATE LENGTH LAYOUT NCOL/WID MODE SD NAME PLEX DISK DISKOFFS LENGTH [COL/]OFF DEVICE MODE SV NAME PLEX VOLNAME NVOLLAYR LENGTH [COL/]OFF AM/NM MODE SC NAME PLEX CACHE DISKOFFS LENGTH [COL/]OFF DEVICE MODE DC NAME PARENTVOL LOGVOL SP NAME SNAPVOL DCO dg remote9960 default default 126000 1032204684.1590.aeneas dm remote996001 AMS_WMS0123_1 auto 2048 1677616896 - dm remote996002 AMS_WMS012_1 auto 2048 1677616896 - dm remote996003 - - - - NODEVICE dm remote996004 AMS_WMS0_1 auto 2048 1677616896 - dm remote996005 - - - - NODEVICE dm remote996006 AMS_WMS0_0 auto 2048 1677616896 - dm remote996007 AMS_WMS0123_0 auto 2048 1258212096 - dm remote996008 AMS_WMS012_0 auto 2048 1258212096 - dm remote996010 AMS_WMS0123_2 auto 2048 2094367888 - dm remote996011 AMS_WMS012_2 auto 2048 2094367888 - dm remote996012 AMS_WMS01_2 auto 2048 2094367888 - dm remote996013 AMS_WMS0_2 auto 2048 2094367888 - v rem-01 - DISABLED ACTIVE 1921724416 SELECT - fsgen pl rem-01-02 rem-01 DISABLED NODEVICE 1921843200 STRIPE 4/64 RW sd remote996004-01 rem-01-02 remote996004 36096 480460800 0/0 AMS_WMS0_1 ENA sd remote996003-01 rem-01-02 remote996003 36096 480460800 1/0 - NDEV sd remote996002-01 rem-01-02 remote996002 36096 480460800 2/0 AMS_WMS012_1 ENA sd remote996001-01 rem-01-02 remote996001 36096 480460800 3/0 AMS_WMS0123_1 ENA v rem-02 - DISABLED ACTIVE 1921724416 SELECT - fsgen pl rem-02-02 rem-02 DISABLED NODEVICE 1921843200 STRIPE 4/64 RW sd remote996004-02 rem-02-02 remote996004 480496896 480460800 0/0 AMS_WMS0_1 ENA sd remote996003-02 rem-02-02 remote996003 480496896 480460800 1/0 - NDEV sd remote996002-02 rem-02-02 remote996002 480496896 480460800 2/0 AMS_WMS012_1 ENA sd remote996001-02 rem-02-02 remote996001 480496896 480460800 3/0 AMS_WMS0123_1 ENA v rem-03 - DISABLED ACTIVE 1000341504 SELECT - fsgen pl rem-03-01 rem-03 DISABLED NODEVICE 1000396800 STRIPE 2/64 RW sd remote996006-02 rem-03-01 remote996006 106826496 419443200 0/0 AMS_WMS0_0 ENA sd remote996008-03 rem-03-01 remote996008 1176381696 80755200 0/419443200 AMS_WMS012_0 ENA sd remote996005-02 rem-03-01 remote996005 106826496 419443200 1/0 - NDEV sd remote996007-03 rem-03-01 remote996007 1176381696 80755200 1/419443200 AMS_WMS0123_0 ENA v rem-04 - DISABLED ACTIVE 1921724416 SELECT - fsgen pl rem-04-02 rem-04 DISABLED NODEVICE 1921843200 STRIPE 4/64 RW sd remote996004-03 rem-04-02 remote996004 960957696 480460800 0/0 AMS_WMS0_1 ENA sd remote996003-03 rem-04-02 remote996003 960957696 480460800 1/0 - NDEV sd remote996002-03 rem-04-02 remote996002 960957696 480460800 2/0 AMS_WMS012_1 ENA sd remote996001-03 rem-04-02 remote996001 960957696 480460800 3/0 AMS_WMS0123_1 ENA v rem-08 - ENABLED ACTIVE 2097152000 SELECT rem-08-01 fsgen pl rem-08-01 rem-08 ENABLED ACTIVE 2097177600 STRIPE 2/64 RW sd remote996008-01 rem-08-01 remote996008 36096 1048588800 0/0 AMS_WMS012_0 ENA sd remote996007-01 rem-08-01 remote996007 36096 1048588800 1/0 AMS_WMS0123_0 ENA v rem-30 - DISABLED ACTIVE 1887436800 SELECT - fsgen pl rem-30-01 rem-30 DISABLED NODEVICE 1887436800 STRIPE 2/64 RW sd remote996006-03 rem-30-01 remote996006 526269696 943718400 0/0 AMS_WMS0_0 ENA sd remote996005-03 rem-30-01 remote996005 526269696 943718400 1/0 - NDEV v rem-40 - ENABLED ACTIVE 2097152000 SELECT rem-40-01 fsgen pl rem-40-01 rem-40 ENABLED ACTIVE 2097254400 STRIPE 4/64 RW sd remote996013-01 rem-40-01 remote996013 36096 524313600 0/0 AMS_WMS0_2 ENA sd remote996012-01 rem-40-01 remote996012 36096 524313600 1/0 AMS_WMS01_2 ENA sd remote996011-01 rem-40-01 remote996011 36096 524313600 2/0 AMS_WMS012_2 ENA sd remote996010-01 rem-40-01 remote996010 36096 524313600 3/0 AMS_WMS0123_2 ENA v rem-41 - ENABLED ACTIVE 2097152000 SELECT rem-41-01 fsgen pl rem-41-01 rem-41 ENABLED ACTIVE 2097254400 STRIPE 4/64 RW sd remote996013-02 rem-41-01 remote996013 524349696 524313600 0/0 AMS_WMS0_2 ENA sd remote996012-02 rem-41-01 remote996012 524349696 524313600 1/0 AMS_WMS01_2 ENA sd remote996011-02 rem-41-01 remote996011 524349696 524313600 2/0 AMS_WMS012_2 ENA sd remote996010-02 rem-41-01 remote996010 524349696 524313600 3/0 AMS_WMS0123_2 ENA v rem-42 - ENABLED ACTIVE 2097152000 SELECT rem-42-01 fsgen pl rem-42-01 rem-42 ENABLED ACTIVE 2097254400 STRIPE 4/64 RW sd remote996013-03 rem-42-01 remote996013 1048663296 524313600 0/0 AMS_WMS0_2 ENA sd remote996012-03 rem-42-01 remote996012 1048663296 524313600 1/0 AMS_WMS01_2 ENA sd remote996011-03 rem-42-01 remote996011 1048663296 524313600 2/0 AMS_WMS012_2 ENA sd remote996010-03 rem-42-01 remote996010 1048663296 524313600 3/0 AMS_WMS0123_2 ENA v rem-43 - ENABLED ACTIVE 2085427200 SELECT rem-43-01 fsgen pl rem-43-01 rem-43 ENABLED ACTIVE 2085427200 STRIPE 4/64 RW sd remote996013-04 rem-43-01 remote996013 1572976896 521356800 0/0 AMS_WMS0_2 ENA sd remote996012-04 rem-43-01 remote996012 1572976896 521356800 1/0 AMS_WMS01_2 ENA sd remote996011-04 rem-43-01 remote996011 1572976896 521356800 2/0 AMS_WMS012_2 ENA sd remote996010-04 rem-43-01 remote996010 1572976896 521356800 3/0 AMS_WMS0123_2 ENA v rimg02 - DISABLED ACTIVE 944793600 SELECT - fsgen pl rimg02-01 rimg02 DISABLED NODEVICE 944793600 STRIPE 4/64 RW sd remote996004-04 rimg02-01 remote996004 1441418496 236198400 0/0 AMS_WMS0_1 ENA sd remote996003-04 rimg02-01 remote996003 1441418496 236198400 1/0 - NDEV sd remote996002-04 rimg02-01 remote996002 1441418496 236198400 2/0 AMS_WMS012_1 ENA sd remote996001-04 rimg02-01 remote996001 1441418496 236198400 3/0 AMS_WMS0123_1 ENA > dxadm Select an operation to perform: 5 Select a removed or failed disk [<disk>,list,q,?] remote996003 VxVM ERROR V-5-2-1985 No devices are available as replacements for remote996003. Select a removed or failed disk [<disk>,list,q,?] remote996005 VxVM ERROR V-5-2-1985 No devices are available as replacements for remote996005. I attemped to reattached the failed disk: bash-2.03# /etc/vx/bin/vxreattach -c remote996003 VxVM vxdisk ERROR V-5-1-537 Device remote996003: Not in the configuration VxVM vxdisk ERROR V-5-1-558 Disk remote996003: Disk not in the configuration bash-2.03# /etc/vx/bin/vxreattach -c remote996005 VxVM vxdisk ERROR V-5-1-537 Device remote996005: Not in the configuration VxVM vxdisk ERROR V-5-1-558 Disk remote996005: Disk not in the configuration bash-2.03# vxdisk clearimport AMS_WMS01_1 VxVM vxdisk ERROR V-5-1-531 Device AMS_WMS01_1: clearimport failed: Disk device is offline I think this part is where I created the two duplicate devices. From this point, I'm going to have to step back and seek out guidance before I cause further problems.Solved2.8KViews0likes6CommentsCreate second RDS using volumes from same disk group in Veritas volume replicator
Hi all, I have one query. I have one disk group with 5 volumes, which are more than 2 TB and are configured for replication. Now i want to add 2 more volumes in the RDS but it will resync all the volumes after adding these 2 volumes, so is it possible to make second RDS for only those 2 new volumes. Also can we use same IP for replication.5.5KViews0likes8Commentsmissing disks and reboot wont solve it
I am very new to veritas. We have AIX 7.1 server using veritas DMP. When I look at the VIO all of the virtual fibre channel adapters are logged in, but on the lpar it is failing to see any disks on fscsi0 and fscsi1. I have been going back and forth with IBM and symantec and cannot get this resolved, so decided to pick your brains here. # lsdev| grep fscsi fscsi0 Available 01-T1-01 FC SCSI I/O Controller Protocol Device fscsi1 Available 02-T1-01 FC SCSI I/O Controller Protocol Device fscsi2 Available 03-T1-01 FC SCSI I/O Controller Protocol Device fscsi3 Available 04-T1-01 FC SCSI I/O Controller Protocol Device fscsi4 Available 05-T1-01 FC SCSI I/O Controller Protocol Device fscsi5 Available 06-T1-01 FC SCSI I/O Controller Protocol Device fscsi6 Available 07-T1-01 FC SCSI I/O Controller Protocol Device fscsi7 Available 08-T1-01 FC SCSI I/O Controller Protocol Device # vxdmpadm listctlr CTLR_NAME ENCLR_TYPE STATE ENCLR_NAME PATH_COUNT ========================================================================= fscsi2 Hitachi_VSP ENABLED hitachi_vsp0 44 fscsi3 Hitachi_VSP ENABLED hitachi_vsp0 44 fscsi4 Hitachi_VSP ENABLED hitachi_vsp0 44 fscsi5 Hitachi_VSP ENABLED hitachi_vsp0 44 fscsi6 Hitachi_VSP ENABLED hitachi_vsp0 44 fscsi7 Hitachi_VSP ENABLED hitachi_vsp0 44 ^ Above you will see that fscsiX seen by OS is not being seen by veritas. How can I force them into veritas? I have already tried rebooting the VIO and LPAR and that doesnt seem to help. FWIW, i deleted the disks that were in defined state. Usually when MPIO is being used and we lose path, deleting the disks and virtual fibrechannel adapter and running cfgmgr solves the issue, but that doesnt seem to help here.Solved