Restores run forever, don't lay files down
Hi,
Here is the background- we backup about 80 NFS shares (off a VNX) through a single media server. The bulk of these gets staged to an Isilon, then copied to tape. We also scrape up daily RMAN-created backups off a seperate VNX and spool those to tape. Single library, 6 LTO5 tape drives, all same media type. Nothing fancy. Been running this setup for years. No changes to buffers or firmware recently.
I recently ran a restore from some of the RMAN tapes which I thought acted wrong in that the KB size written was going well beyond what the actual source files were according to the browse/restpore file list. The job ran and ran, never laid even a byte back down, so I killed it thinking it was a fluke. I then tried to restore a different set of files from that same time period, same results, runs forever, nothing gets written to the destination, whether same or redirected. I then went and randomly chose some other data from other NFS mounts and those all worked fine, in fact I cannot reproduce the problem except on the Oracle-created backups. The only difference I can see is that the RMAN stuff is going direct to tape, no Isilon staging involved. We have restored this same RMAN data before using the same process, but it's been a few months since the last time. Anyhow, I've never encountered anything like this where the restore job runs excessively long, appears to be 'working' (there is a process chewing away at CPU and memory, and the tape drive is spinning, yet nothing gets restored. What could be going on?
thanks
Have seen behaviour like this with unlimited fragment size, in that case the problem was that the system had read the whole tape before extracting the file(s).
The solution/migation was to use a set fragment size, unfortunately this only solve/migate the this for future backups unless you duplicate the earlier/current backups after changing the fragment size.