cancel
Showing results for 
Search instead for 
Did you mean: 

Exchange snapshot takes 2 hours

Marianne
Level 6
Partner    VIP    Accredited Certified

Environment:

Exchange 2010 Standalone server
NBU 7.0.1 on master/media server and client
W2008 R2 SP1

Exchange policy with 3 Information Stores in Backup Selection with multiplexing and multistreaming enabled.

Backup starts and snapshots are created on Exchange Client.

bpfis log:

 

12:00:08.142 [9876.17364] <4> bpfis: INF - Preparing freeze of NEW_STREAM0 using snapshot method VSS_Writer.
12:00:08.142 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_NEW_STREAM0_9876
12:00:08.157 [9876.17364] <4> bpfis: INF - Preparing freeze of Microsoft Information Store:\Archive DB3\ using snapshot method VSS_Writer.
12:00:08.157 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_Microsoft Information Store_Archive DB3__9876
12:00:34.335 [9876.17364] <4> bpfis: INF - Preparing freeze of NEW_STREAM1 using snapshot method VSS_Writer.
12:00:34.335 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_NEW_STREAM1_9876
12:00:34.335 [9876.17364] <4> bpfis: INF - Preparing freeze of Microsoft Information Store:\Archive DB9\ using snapshot method VSS_Writer.
12:00:34.335 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_Microsoft Information Store_Archive DB9__9876
12:00:36.363 [9876.17364] <4> bpfis: INF - Preparing freeze of NEW_STREAM2 using snapshot method VSS_Writer.
12:00:36.363 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_NEW_STREAM2_9876
12:00:36.363 [9876.17364] <4> bpfis: INF - Preparing freeze of Microsoft Information Store:\Archive DB10\ using snapshot method VSS_Writer.
12:00:36.363 [9876.17364] <4> bpfis: INF - Created mount point C:\Program Files\Veritas\NetBackup\temp\_vrts_frzn_img_Microsoft Information Store_Archive DB10__9876
12:00:38.438 [9876.17364] <4> bpfis: TRV - snapshot preparation successful
13:50:33.580 [9876.17364] <2> bpfis main: snapshot creation successful
 
 
Have a look at the last two lines - just short of 2 hours between 'snapshot preparation' and 'snapshot creation'.
bpbkar log starts at 13:56.
 
Unfortunately nothing in Event Viewer Application log as the log got recreated last night (first entry about 23:30).
 
Can anyone tell me what is happening in the 2 hours?
Does this delay include database validation?
 
Thanks!
 
 
22 REPLIES 22

jandersen1
Level 4
Partner Accredited

Hi,
A follow up on my previous post: My customer verified that the MS hotfix was already installed (included in an update).

And then a little more info which might be useful.

Although we are running a DAG with two servers the symptoms are pretty much the same as Marianne's.
However we got a decent performance by disabling the concistency check and using the passive copy for backup of the 4TB (8 databases and 2 public folders). All completed in 5 hours 30 minutes. Backup is to an Windows MSDP and with client side dedupe disabled (I didn't yet try it with client side dedupe enabled).

To me the symptomps looked like som kind of contention (maybe storage or filesystem) when the snapshot verification was taking place. When it was disabled (which you can't do as it is a standalone) all streams started after few minutes instead of 2-3 hours.

--jakob;

 



 

Marianne
Level 6
Partner    VIP    Accredited Certified

Hi All

Thank you for all the advice, pointers, assistance, etc.. etc...

I think Jakob is correct - my gut feel is that we are dealing with resource contention on the Exhange Server.

This is what we did yesterday afternoon:

Uninstalled NBU client software and reinstalled. 

We kicked off a backup of a single IS.

This time, the Evt Application log looked totally different!
No more  "Cmdlet failed" Event Errors' and VSS snapshot process could be seen:

 

Level Date and Time Source Event ID Task Category
Information 2013/05/21 05:21:00 PM MSExchangeIS 9622 Exchange VSS Writer Exchange VSS Writer (instance 09f024e6-fad8-4620-8922-7be25f4de44c:3) has processed the post-snapshot event successfully.
Information 2013/05/21 05:20:46 PM MSExchangeIS 9612 Exchange VSS Writer Exchange VSS Writer (instance 09f024e6-fad8-4620-8922-7be25f4de44c:3) has thawed the database(s) successfully.
Information 2013/05/21 05:20:46 PM ESE 2003 ShadowCopy "Information Store (5164) Shadow copy instance 3 freeze ended.

For more information, click http://www.microsoft.com/contentredirect.asp."
Information 2013/05/21 05:20:45 PM MSExchangeIS 9610 Exchange VSS Writer Exchange VSS Writer (instance 09f024e6-fad8-4620-8922-7be25f4de44c:3) has frozen the database(s) successfully.
Information 2013/05/21 05:20:45 PM ESE 2001 ShadowCopy "Information Store (5164) DB1: Shadow copy instance 3 freeze started.

For more information, click http://www.microsoft.com/contentredirect.asp."
Information 2013/05/21 05:20:45 PM ESE 2001 ShadowCopy "Information Store (5164) Shadow copy instance 3 freeze started.

For more information, click http://www.microsoft.com/contentredirect.asp."
Information 2013/05/21 05:20:45 PM MSExchangeIS 9608 Exchange VSS Writer Exchange VSS Writer (instance 09f024e6-fad8-4620-8922-7be25f4de44c:3) has prepared for Snapshot successfully.
Information 2013/05/21 05:20:45 PM MSExchangeIS 9811 Exchange VSS Writer Exchange VSS Writer (instance 3) has successfully prepared the database engine for a full or copy backup of database 'DB1'.
Information 2013/05/21 05:20:45 PM ESE 2005 ShadowCopy "Information Store (5164) Shadow copy instance 3 starting. This will be a Full shadow copy.

For more information, click http://www.microsoft.com/contentredirect.asp."
Information 2013/05/21 05:19:13 PM MSExchangeIS 9606 Exchange VSS Writer Exchange VSS Writer (instance 09f024e6-fad8-4620-8922-7be25f4de44c) has prepared for backup successfully. 

 

NBU bpfis was at this point still saying 'successful snapshot preparation'.

We can only assume that a consistency check was running at this point, as 'snapshot created' message only came through an hour later and backup started writing.

So, it seems that the consistency check is running a bit faster when there is only a single DB to check, but it still does not help to get 15 DBs backed up in one night. Disabling the consistency also made no difference as it seemed to run in any case (probably because it is standalone?). 
Again - this was no problem when there were 3 large stores on the previous Exchange server....

 

jandersen1
Level 4
Partner Accredited

Hi,

Oups I forgot to mention the probably most important change we made in our setup.
We limited the number of simultanious streams for each of the DAG members to three and found that it run must faster.


In your case it should be enough set the maximum number of streams per policy to a low number (maybe 3 or 4). For a DAG with an even distribution of MBs across the DAG members we found that it was necesarry to do it at NBU client level (client attributes on master server properties). Otherwise the first member could easily takes up all the streams for the policy and you were back to square one.

You need to know a little more about the underlying storage array used for the Exchange databases. If it is using the same total number of spindles for the LUNs as before then you are definately creating contention by running 5 x the number of read streams for the same number of harddisk readheads.

--jakob;