NBU 5220 Performance
We recently deployed a 5220 appliance into our environment as it was to be the savior in our battle against a backup window we were no longer able to meet. When we finally got it online and into our NBU environment the initial performance was great. The area we were to most benefit was with VMware backups. The data stores mounted directly to the appliance would allow direct access to the snapshots for a fast and efficient backup. Before this, we were performing client side backups so the impact on the hosts every night was significant as we tried backing up 600+ vms. The plan was to be able to move all Dev and test off-host backups to the middle of the day as the performance impact was minimal and the end result was an increased window to complete things. As we started with noon-time backups deduplication rates were high and so were the speeds. However, this performance gain was short lived, as we began increasing the load we suddenly saw performance drop to a point of concern. Backups were no longer speedy, 3500KB per second to 10,000KB. There are some that might pop to 24,000KB, but in a sample size of 15 as I write this, only one is showing 24,000. Now, I do have a few ideas. 1. We have a 72TB appliance, therefore, 2 disk trays, during the backups, the one disk tray is going crazy, all the lights are flashing and you can really see that it is working. However, the second tray is doing nothing. While you might see a blink here or there, it is almost nothing compared to the other disk tray. Is this to be expected? When we looked at the disk configuration, it shows concat, is this normal? 2. Too much data at once and we are simply burying the appliance. In reality, what sort of performance should I be able to expect from the appliance? 3. Relating to number 2, since we only have the one appliance right now while we wait to get the remote appliance in place, we are duping off to tape. This is running at the same time as a backup, so this means at the same time the appliance is writing a lot of data, it is also reading it back to tape. 4. We are overloading the data store so that read speed is bad from source to destination. We have fewer hosts,therefore, if we limit the jobs per host we limit the number of machines backing up at once (obviously). This means that backups take way long, so we removed the limit per host and just set a limit per data store. As we are new to it all, I am not sure what impact where, but again, I am trying to list any and all ideas from the start. 5. The appliance does not support multi-pathing, therefore, we only have a single path to the disk. Beyond that I am not sure, but this is something that doesnt help with the showcasing of the appliances to management at the moment. However, given the initial performance I am confident we can get back there.Solved3.2KViews4likes50Commentsduplications from wrong media server
I have two 5200 appliances (2.0.1) one is in our primary data center and one is at our DR site connected by a 43mb pipe. My primary appliance has two LTO tape libraries directly attached via fiber that are used for doing monthly backups. I am using SLP's for the back up and duplications. Every weekend we do a full backup (1.3tb) that is backed up to my primary appliance and then duplicated to my DR appliance. During this backup i notice that my media server that is being used is primary > primary (for backup) and DR > DR (for duplications). This is perfectly fine. Now heres the trouble. When i mix a duplication to tape in this process the system trys to load balance the duplication between my media servers. Meaning: One duplication job will show the primary as the media server for the tape backup and one job will show the DR appliance. This would all be ok if the DR appliance was on my local LAN but it is not and therefore any duplication using it as the media server to tape can take over 24 hours to complete. In the SLP for my monthlys I have removed the DR appliance as a duplication so i dont know why it keeps responding. Example: i had a 245gb exchange DB duplication that was using my DR appliance as the media server and it had been running for 48hours and was only 60% complete. I noticed the job and cancled it. When it retried it picked up my primary appliance as the media server and in 25min it was complete (over fiber channel). I opened a ticket with support 2 months ago but we have yet to figure out the deal. This hasnt been that big of an issue until another group wanted to replicate some VTL's to the DR and realized that the bandwith is 100% consumed all the time by netbackup. i have been tinkering with this for months and cannot figure out what i have configured wrong so i want to consult the almight Community for answers. LOL! To me it sounds like the media servers are setup so that they can load balance but i cant find were to disable that.Solved1.9KViews4likes6CommentsWhen too much process on appliance, randomly some of database backup fails at the middle of process
Hello, Firstly, this is a big company, nearly 20-25 TB data is backedup as weekly full backup. I think the problem occurs when i have duplication and backup processes at the same time. Randomly some db backup fails with 6 errors at the middle of job. Sometimes i make rebase is off and these errors are lower than to rebaseon. After a fail, i retry to job and it mostly finishes successfully. i dont exactly know which logs are needed. i try to explain my envoirement. Master -> 7.5.0.7, windows 2008r2 Media -> 5220 appliance, 2.5.4 Clients -> windows 2008r2, sql2008 i found some link about the issue, but i dont know exactly these changes work or not. http://www.symantec.com/connect/forums/interrupt-when-backup-large-sql-databases http://www.symantec.com/connect/forums/5220-limits?page=0#comment-10092541 Thanks a lotSolved1.4KViews3likes6CommentsCancel retrying (AIR) Import job on 5200 master
I have configured one of my 5200's as a master server and put it in our dr site. Im am using AIR and import SLP's to replicate the data to the DR appliance. Since I upgraded my datacenters netbackup to 7.5 imports for my cataloge backups are failing. Support has already confirmed this is a bug and will be fixed in the next release (7.5) of the appliance. So I have stopped replication of the cataloge backups. But... My issue is the original failing imports will not stop retrying. This has been going on for weeks and i cant seem to figure out how to stop them on the appliance. I've ran commands in the past for my windows NBU master serve to get jobs to start but this is my first go around with an appliance. Does any one know how to kill these jobs? Each one shows the same info in them and they have been retrying for weeks now. Thanks.Solved1.7KViews3likes3CommentsNetBackup_7_Deduplication_Sizing Tool-Basic-3.2.xlsx
I have found a link to this excel on two posts in this forum. As this excel refers to some external data or calculation, I cannot use it In addition, it is not adjuted to 2.6 (7.6.0.1) Who owns the source? Can I get it? RegardsSolved831Views2likes3CommentsBuffer Settings for Duplication to LTO 5 Tape
We're using a Netbackup 5230 Appliance connected via 8Gb fibre to an HP MSL 8096 LTO 5 tape library with 4 drives. I've read many posts recommending leaving the buffer size as default: SIZE_DATA_BUFFERS : 262144 (Default) SIZE_DATA_BUFFERS_DISK : 262144 (Default) SIZE_DATA_BUFFERS_FT : 262144 (Default) This is how they are currently set, and when monitoring backups to disk, I see the following in the log: 01/29/2014 19:49:09 - Info bptm (pid=3752) using 262144 data buffer size 01/29/2014 19:49:09 - Info bptm (pid=3752) using 30 data buffers Which is expected, however, when duplicating to tape, the logs indicate the following: 01/31/2014 17:00:15 - Info bptm (pid=30607) using 65536 data buffer size 01/31/2014 17:00:15 - Info bptm (pid=30607) using 30 data buffers Can someone clarify why it would be set this way and is this as expected? ThanksSolved2.6KViews2likes13CommentsNetBackup Appliance 5220 Slow Rehydration to tape
Hi Everyone, I know there have been lots of comments surrounding slow rehydration to tape from the appliances but mine is just to slow and the appliance can't keep up with the demand. Master server = Windows 2008 R2 SP1 NetBackup 7.5.0.3 Appliances 5220 = NetBackup 7.5.0.4 (patch 2.5) Tape drives linked to appliance at 4GB over fibre ibm 3592 drives capable of over 100mb a second native without compression over 200mb a second when compressing. We have an SLP to backup our Exchange servers to a 5220 appliance then replicated to another 5220 on a different site but same netbackup domain and then rehydrate to tape from there. We have 12 tape drives avalaible but currently the storage unit is set to allow 4 concurrent write drives for the duplication to tape. The exchange servers are the passive nodes and we are using client side dedupe on them as its faster than using the appliance dedupe as there are 6 servers totaling in about 13TB of data. All buffers are set to appliance defaults at 256kb (262144) and using 30 buffers. we currently use this on our master server catalog backup and get over 120MBps Currently tape speeds on rehydration are around 15 -25MBps topping out across all 4 streams at 75-85MBps. This speed is appauling when you have 13TB a night to get to tape. I have tried running a single stream to tape and it still can only produce around 50-60MBps Has anyone had any luck improving this?946Views2likes6CommentsNetbackup Appliance deduplication/re-hydration problem
Netbackup Appliance deduplication/re-hydration problem Netbackup Master Server Version: 7.5.0.7 Netbackup Appliance 2530, Version 2.5.4 Netbackup Appliance Media Server Version 7.5.0.7 Backup Type: Exchange DB Backup (DAG) with GRT Client OS/Version: Windows Server 2008 RT, Netbackup Help! One of our Netbackup appliance hosts Exchange Database backup (DAG) with GRT... Backups do run with good speed and without any issues during normal circumstances... These backups are set to duplicate to tape (managed via SLP), everyweek... As soon as the SLP starts to copy the data from MSDP to tape, the backups' speed becomes haywire and runs extremely slowly.. As soon as I kill off the duplication job, the backups revert to normal again... Another problem is that, whenever the duplication run the exchange backups fail with error code 24 or 42 (Network Write error or Network Read error).. This doesnt happen during normal backups... Please help on this. Also, it would be highly helpful if someone could explain de-duplication/dehydration workflow and duplication to tape workflow (via SLP)...Solved1.3KViews1like6Comments5230 Appliance Send Mail Notification
Hello Team, We used appliance for backup i wan to configure send mail notification on master server admin guı not Ops Center? This can be possible i successd for appliance web page and sent it e-mail for users but same time i tried to send e-mail notification from appliance master server gui which does not work ..... Could you please inform me about "send mail notification for appliance" BR KaccoSolved2.6KViews1like2CommentsUnderstanding Accelerator (from detailed Status view)
Hi all, I am in need of confirmation or explanation on some logging I see in the detailed status. When I see this in the detailed status (for an Incremental backup): 04/15/2015 07:02:43 - Info bpbkar (pid=98149) accelerator sent 1115317760 bytes out of 5630247936 bytes to server, optimization 80.2% Does this mean the Accelerator has found 5630247936 bytes of new/changed data on the client of which 1115317760 bytes is new data to the backup appliance? And how does this related to: 04/15/2015 07:02:45 - Info netbackup (pid=164455) StorageServer=PureDisk:netbackup; Report=PDDO Stats for (netbackup): scanned: 5498334 KB, CR sent: 268206 KB, CR sent over FC: 0 KB, dedup: 95.1%, cache disabled I see that the client-side-dedup has found scanned: 5498334 KB of new/changed data of which sent: 268206 KB is sent to the Backup Appliance as this is new unique data. Please correct me where I am wrong or comfirm I am understanding this correctly. Many thanks Regards, Ronald1.2KViews1like6Comments