05-11-2010 04:16 AM
05-11-2010 06:29 AM
05-11-2010 08:49 AM
09-12-2010 11:17 AM
Hey guys,
Has anyone achieved this goal?
I've read (don't remember where) that BE Dedup is based on the source of the data, so this strategy suggested by Collin wouldn't work.
For example, if a have an 100MB file in a server, backed up via source dedup, and then i try to backup the same 100MB file in a second server, it wouldn't transmit only unique blocks, as they are from different sources/servers.
I imagine GLOBAL DEDUP (the oposite of the example) would happen only in PureDisk (full version).
Any thoughts? I basically have a 50GB database, trying to backup it via 4Mbps WAN (first backup should run for 70hrs, too much for an instable link).
Thanks a lot in advance!
09-12-2010 08:21 PM
Rafael,
On client side dedupe, the client will scan all data to be backed up, the data broken up into segments and an MD5 (like) hash created. This data is then sent to the BE server with the dedupe option installed and configured.
From there the BE server responds back to the client with what segments to send and what it already has.
So Colin's scenario would work.
09-13-2010 11:28 AM
Hi Teiva,
Thanks for your reply. Actually what we've seen is that the job is copying all the data again. Here's my scenario (detailed):
-ServerA: BE2010 R2 Media Server + Dedup Folder
-ServerB: SQL 2008 R2 (50GB database)
-ServerC: SQL 2008 R2 (50GB almost identical database.. gap of 1 day, but the database had really few updates).
ServerA and ServerB are near (1Gbps connection), and ServerC is far away (4Mbps connection).
My steps:
1)Backup ServerB database to the Dedup Device - total time = 2hs, dedup ratio of 1.1:1 (first copy).
2)Backup ServerB database again - total time = 10mins, dedup 314:1 (so dedup is working, and direct access too)
3)Backup ServerC database - copied 2GB in 3hrs, so source dedup is not working.
I've cancelled the job to see if it was working with RADA, and it was (the warning stating that source dedup was not used didn't appeared). If I block port 10000 and re-run the job, it shows me the warning.
Another curious fact: If I run the job again, it will go fast to the 2GB (1GB/min), and then slow down again. So we can conclude that the 2GB that were copied from ServerC (first job) in fact were stored, but the rest of the data is not being compared with the MD5 hashs generated from ServerB backup.
Any ideias?
Thanks a lot for your help, and sorry the bad english. :)
Rafael
09-13-2010 12:36 PM
Guys,
I've made another test with a file, just to be sure that it wasn't an behavior caused by backing up an SQL Database.
1)Copied an 100MB file to the local server (ServerB) and remote server (ServerC).
2)Run a backup of the ServerB (40 seconds).
3)Run a backup of the ServerC with source dedup (5 minutes) - Source dedup didn't compared the blocks with the data from the other server.
4)Run again a backup of the ServerC with source dedup (20 seconds) - Source dedup DID work, but clearly was compared with the hashes from ServerC (step 3) and not ServerB (step 2)
In a few words: No source dedup from different machines.
12-09-2010 01:31 PM
I am looking to setup a Dedup server on my network where the Media server will backup files from 10 different servers all over a WAN connection, in remote locations. One of which is an Exchange server with 300GB of databases and other SQl databases on another server, the rest are file servers.
I understand about pre-staging the file servers on a USB drive from the source using an identical file structure and backing it up onto the destination media server using dedup. But how can I pre-stage the Exchange databases and SQL databases? Will these databases also get pre-staged on a USB drive using the same identical directory structure from the source exchange server? Or does it not work the same for live exchange databases? Same for SQL databases, will the SQL databases need to be backed up on a USB with the same directory structure?
Thanks, this post has been very helpful.
Rob