DRL not working on mirrored volumes in VVR - RVG (Mirrored volume doing full resync of plexes)
I think I am hitting a major issue here with a mirrored volume in RVG. SRL is supposed to provide the DRL functionality . Hence DRL logging is explicitly disabled when a volume is added to RVG. However my testing shows that DRL is not working and in the case of a mirror plex out of sync due to a server crash etc, full resync of mirror plexes is happening. (not just the dirty regions). Here is a quick and easy way to recreate the issue: My configuration: Infoscale 8 Redhat 8.7 I have a mirrored volume sourcevol2 (2 plexes) which I created like below: #vxassist -g dg1 make sourcevol2 1g logtype=dco drl=on dcoversion=20ndcomirror=1 regionsz=256 init=active #vxassist -b -g dg1 mirror sourcevol2 I wait for the synchronization to complete #/opt/VRTS/bin/mkfs -t vxfs -o nomaxlink /dev/vx/rdsk/dg1/sourcevol2 # mount /dev/vx/dsk/dg1/sourcevol2 /sourcevol2 I create SRL as below: #vxassist -g dg1 make dg1_srl 1g layout=concat init=active I create primary rvg as below: #vradmin -g dg1 createpri dg1_rvg sourcevol2 dg1_srl Verified dcm in dco flag is on. #vxprint -g dg1 -VPl dg1_rvg |grep flag flags: closed primary enabled attached bulktransfer dcm_in_dco Added secondary #vradmin -g dg1 addsec dg1_rvg primarynode1 primarynode2 Started initial replication #vradmin -g dg1 -a startrep dg11_rvg primarynode2 Verified replication is uptodate #vxrlink -g dg1 -T status rlk_primarynode2_dg1_rvg VxVM VVR vxrlink INFO V-5-1-4467 Rlink rlk_primarynode2_dg1_rvg is up to date Here is the actual scenario to simulate mirror plexes out of sync : On primary: Run a DD command to put some IO on sourcevol2 #dd if=/dev/zero of=/sourcevol2/8krandomreads.0.0 bs=512 count=1000 oflag=direct In another terminal , force stop the sourcevol2 while dd is going on. #vxvol -g dg1 -f stop sourcevol2 #umount /sourcevol2 Start the sourcevol2 #vxvol -g dg1 start sourcevol2 #vxtask -g dg1 list -l Task: 160 RUNNING Type: RDWRBACK Operation: VOLSTART Vol sourcevol2 Dg dg1 Even though I only changed only few regions on the sourcevol2 (sequential writes of 512b), the volume goes through a full plex resync (as indicated by the time to start the volume). Summary: DRL on a Volume added to an RVG is not working . Hence mirrored volumes are going through a full plex resync as opposed to only resync of dirty regions.1.7KViews0likes5CommentsDoes Infoscale Storage (VVR) support cascaded space-optimized snapshot?
Configuration: Infoscale Storage 8.0 on Linux Infoscale storage foundation supports cascaded snapshot using vxsnap infrontof= to do cascaded snapshots Infoscale storage (with Volume replicator) documentation doesn't describe cascaded snapshot. I checked manpage for vxrvg. Does not have an infront of attribute. Does that mean cascaded space-optimized snapshots are not supported/permitted on RVG?1.2KViews0likes2CommentsDoes VVR support cascaded space-optimized snapshot?
Configuration: Infoscale Storage 8.0 on Linux Infoscale storage foundation supports cascaded snapshot using vxsnap infrontof= to do cascaded snapshots Infoscale storage (with Volume replicator) documentation doesn't describe cascaded snapshot. I checked manpage for vxrvg. Does not have an infront of attribute. Does that mean cascaded space-optimized snapshots are not supported/permitted on RVG?728Views0likes0CommentsWhen life give you lemons, make cliff notes!
At Symantec, we are always looking for ways to empower you to do more with less! What better ways to do that than create cliff notes to quickly guide you through some Storage Foundation (SF) and Veritas Cluster Server (VCS) tasks! Thanks to the Symantec Education team, attached are four quick reference guides! Go ahead and use them and you will have plenty of time to make that4.5KViews0likes7CommentsMaxuproc not get updated even after reboot
Hi, Got a update to change the "maxuproc for wt82369 by 1.5 times" , While verifying we make necessary modification on the Global (wt81958). Normally there is a relation between max_nprocs value and maxuproc value. FYI.. maxuprc = max_nprocs – reserved_procs (default is 5) In this case we modified the max_nprocs value from 30000 to 50000 FYI.. [root@wt81958 GLOBAL] /etc # cat /etc/system | grep max_nprocs set max_nprocs=50000 After the global zone reboot the value is not updated while we hit sysdef [root@wt81958 GLOBAL] /root # sysdef | grep processes 30000 maximum number of processes (v.v_proc) 29995 maximum processes per user id (v.v_maxup) Can anyone please assist us if any thing we missed in this to make the necessary changes to replicate. Awaiting for your valuable suggestions. Thanks, senthilsamSolved3.2KViews0likes3CommentsAlmost done with installation, not completed yet
Oracle Automatic Service Request (ASR) completed. Initiated walk through and transfer of system setup from Oracle to Symantec. Registering the support services still seeing issues. The Support ID is assigned to someone at Verisign. Multiple emails, phone calls to resolve the problem. Need to follow up with Oracle support next week.765Views0likes2CommentsAdding New Node Veritas Cluster Server with different hardware specification
Dear Experts, I need your suggestion on the below: Currently we have Two Node Veritas Cluster 6.2 running Windows 2008 R2 hosted on HPE DL380 G7 Servers. We are planning to refresh the hardware and want to move all workloads to new HPE DL380 G9/G10 Servers with Veritas Cluster 6.2 being deployed on Windows 2008 R2. It will only hardware refresh without any Application OR OS Upgrade. Currently Oracle 10gR2 is configured in Failover cluster mode. Application binaries are installed in C:\ drives on all cluster nodes. Would like to know whether I can deploy New VCS 6.2 node on New HPE DL380 G9/G10 Server and Add to existing cluster? If possible, what is the way around. OR this will not work? I tried to search articles, but no luck. Since the hardware architecture will be different, what will be the consequences when we do failover manually OR if we shutdown the Resource Group and start on Newly deployed server? Appreciate you feedback, answers, and any ideas with new approach. Thanks RaneSolved2.5KViews0likes5Commentsvxlicrep ERROR V-21-3-1015 Failed to prepare report for key
Dear all, we got a INFOSCALE FOUNDATION LNX 1 CORE ONPREMISE STANDARD PERPETUAL LICENSE CORPORATE. I have installed key using the vxlicinst -k <key> command. But when I want to check it using vxlicrep I'm getting this error for the given key: vxlicrep ERROR V-21-3-1015 Failed to prepare report for key = <key> We have Veritas Volume Manager 5.1 (VRTSvxvm-5.1.100.000-SP1_RHEL5 and VRTSvlic-3.02.51.010-0) running on RHEL 5.7 on 64 bits. I've read that the next step is to run vxkeyless set NONE, but I'm afraid to run this until I cannot see the license reported correctly by vxlicrep. What can I do to fix it? Thank you in advance. Kind regards, Laszlo4.2KViews0likes7Comments