Re: [Veritas-bu] NDMP and Millions of Files

2010-06-21 Thread Bahadir Kiziltan
Try NetApp NDMP streamer, which allows you to leverage the deduplication in
NDMP backups.

You need at least PureDisk 6.6.x + NBU 6.5.4 with EEB.

On Thu, Jun 17, 2010 at 5:58 PM, ccosta@gmail.com wrote:

 Good news is that with 7.01 coming out in a month or so will allow NBU to
 multi-stream NDMP data to a single tape drive.

 This will/may alleviate some of the performance issues many of you
 experience each day. However I am not sure of any limitations of this
 feature patch may have.

 Chris

 Sent from my Verizon Wireless BlackBerry

 -Original Message-
 From: Ambrose, Monte mambr...@qualcomm.com
 Date: Thu, 17 Jun 2010 14:38:41
 To: Jonathan Dyckjd...@bank-banque-canada.ca; Jeff Cleverley
 jeff.clever...@avagotech.com; rusty.ma...@sungard.com
 rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.eduveritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 You could also use NetBackups Snap Mirror to tape.  You would need a NetApp
 snap mirror license.

 The Pros
 It uses snapmirror and sends the data off to tape.  It is a RAW volume
 backup and is extremely fast - 10X in many cases.
 It can be fully configured in NetBackup

 The Cons
 It backs up the entire volume - so if you have a 1TB volume and only 400GB
 are used it will backup 1tb
 You cant restore a single file or dir.  You have to restore the entire
 thing.
 You cant mix with incremental backups.

 Monte

 -Original Message-
 From: veritas-bu-boun...@mailman.eng.auburn.edu [mailto:
 veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jonathan Dyck
 Sent: Tuesday, June 15, 2010 7:21 AM
 To: Jeff Cleverley; rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 We do our NetApp backups (old GF940 metro cluster) with a combination of
 snapshots (which are available at both sites) and NDMP here. Our NDMP
 obviously holds all the long retention data.

 The data size isn't huge compared to some (6.5TB,  ~40M files), but we've
 had to resort to multiple policies and multi-streaming to back it up in a
 reasonable amount of time (less than 60 hours for a full on the wknd).  The
 way it works is:

 PolicyA (vol1): explicitly lists 37 different paths for the backup
 selection list,  we've empirically determined these are the smaller
 folders
 PolicyB (vol1-long): explicitly lists 16 paths for the backup selection
 list,  we've determined these are the large folders
 PolicyC (vol1-Catch-missed-directory): we've mounted the root of vol1 on a
 Linux host,  and we back it up via NFS, excluding the 37+16 paths defined
 above.  If this policy's full backup every gets too large (over 10GB or so),
  we review the contents and add new paths to PolicyA or PolicyB as
 necessary.  This is necessary because you can't specify wildcards on NDMP
 backups (discussed in this forum several times I believe).

 We repeat the above process for vol2.

 The above backup data sits on a deduped VTL for 2 months, and then the data
 that is held longer than that is duplicated to tape for long-term storage
 and expired off the VTL.  As we run 5 streams concurrently, the throughput
 is decent, but we peg out the CPU on the NetApp frequently during the backup
 window,  which is a concern.

 HTH...
 Jon



 -Original Message-
 From: veritas-bu-boun...@mailman.eng.auburn.edu [mailto:
 veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jeff Cleverley
 Sent: June 14, 2010 5:57 PM
 To: rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 Rusty,

 If you have a way to use Snapvault to backup to another location, I
 would use it.  We have a number of file systems like what you have.  I
 tried NDMP over TCP and NFS backups using dedicated snapshots mounted
 on a client.  Both used a dedicated 10G network.  We basically overran
 our 6030 filer.  We could have jumped through a lot of hoops and split
 backups over multiple weekends, etc, but we decided it wasn't worth
 it.  I haven't tried Flash Backup for a while but it didn't really buy
 us much on what we tried to do with it.  It may work better now.

 We backup everything (~200 TB) to NearStores in another building.  We
 use SnapVault instead of SnapMirror.  We can still revert our
 destination volumes to primary r/w file systems if we need to.  We
 don't have the requirement to send tapes off site.  If you do, you
 could still make the tapes from your secondary filer.

 Using NFS to tape will give you check points and you can run multiple
 streams to each tape drive.  If you are not hitting the filer
 throughput limits this may work for you.

 Jeff

 On Mon, Jun 14, 2010 at 3:24 PM,  rusty.ma...@sungard.com wrote:
 
  We have a NetApp filer that has a few TB of data made up largely of
 millions
  of small files (about 30 million or so) and we are using several NDMP
  policies to back up this data. The two main problems are length of time

Re: [Veritas-bu] NDMP and Millions of Files

2010-06-21 Thread Ed Wilts
On Mon, Jun 21, 2010 at 2:19 AM, Bahadir Kiziltan 
bahadir.kizil...@gmail.com wrote:

 Try NetApp NDMP streamer, which allows you to leverage the deduplication in
 NDMP backups.

 You need at least PureDisk 6.6.x + NBU 6.5.4 with EEB.


The PureDisk requirement is the killer. I've got several 20TB applications
with hundreds of millions of files.  Purchasing PureDisk licenses for that
much data is prohibitively expensive.

As long as Symantec insists on a per-TB license for PureDisk, we will
continue to make as little use of it as we can.

   .../Ed

Ed Wilts, RHCE, BCFP, BCSD, SCSP, SCSE
ewi...@ewilts.org
Linkedin http://www.linkedin.com/in/ewilts




 On Thu, Jun 17, 2010 at 5:58 PM, ccosta@gmail.com wrote:

 Good news is that with 7.01 coming out in a month or so will allow NBU to
 multi-stream NDMP data to a single tape drive.

 This will/may alleviate some of the performance issues many of you
 experience each day. However I am not sure of any limitations of this
 feature patch may have.

 Chris

 Sent from my Verizon Wireless BlackBerry

 -Original Message-
 From: Ambrose, Monte mambr...@qualcomm.com
 Date: Thu, 17 Jun 2010 14:38:41
 To: Jonathan Dyckjd...@bank-banque-canada.ca; Jeff Cleverley
 jeff.clever...@avagotech.com; rusty.ma...@sungard.com
 rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.eduveritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 You could also use NetBackups Snap Mirror to tape.  You would need a
 NetApp snap mirror license.

 The Pros
 It uses snapmirror and sends the data off to tape.  It is a RAW volume
 backup and is extremely fast - 10X in many cases.
 It can be fully configured in NetBackup

 The Cons
 It backs up the entire volume - so if you have a 1TB volume and only 400GB
 are used it will backup 1tb
 You cant restore a single file or dir.  You have to restore the entire
 thing.
 You cant mix with incremental backups.

 Monte

 -Original Message-
 From: veritas-bu-boun...@mailman.eng.auburn.edu [mailto:
 veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jonathan Dyck
 Sent: Tuesday, June 15, 2010 7:21 AM
 To: Jeff Cleverley; rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 We do our NetApp backups (old GF940 metro cluster) with a combination of
 snapshots (which are available at both sites) and NDMP here. Our NDMP
 obviously holds all the long retention data.

 The data size isn't huge compared to some (6.5TB,  ~40M files), but we've
 had to resort to multiple policies and multi-streaming to back it up in a
 reasonable amount of time (less than 60 hours for a full on the wknd).  The
 way it works is:

 PolicyA (vol1): explicitly lists 37 different paths for the backup
 selection list,  we've empirically determined these are the smaller
 folders
 PolicyB (vol1-long): explicitly lists 16 paths for the backup selection
 list,  we've determined these are the large folders
 PolicyC (vol1-Catch-missed-directory): we've mounted the root of vol1 on a
 Linux host,  and we back it up via NFS, excluding the 37+16 paths defined
 above.  If this policy's full backup every gets too large (over 10GB or so),
  we review the contents and add new paths to PolicyA or PolicyB as
 necessary.  This is necessary because you can't specify wildcards on NDMP
 backups (discussed in this forum several times I believe).

 We repeat the above process for vol2.

 The above backup data sits on a deduped VTL for 2 months, and then the
 data that is held longer than that is duplicated to tape for long-term
 storage and expired off the VTL.  As we run 5 streams concurrently, the
 throughput is decent, but we peg out the CPU on the NetApp frequently during
 the backup window,  which is a concern.

 HTH...
 Jon



 -Original Message-
 From: veritas-bu-boun...@mailman.eng.auburn.edu [mailto:
 veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jeff Cleverley
 Sent: June 14, 2010 5:57 PM
 To: rusty.ma...@sungard.com
 Cc: veritas-bu@mailman.eng.auburn.edu
 Subject: Re: [Veritas-bu] NDMP and Millions of Files

 Rusty,

 If you have a way to use Snapvault to backup to another location, I
 would use it.  We have a number of file systems like what you have.  I
 tried NDMP over TCP and NFS backups using dedicated snapshots mounted
 on a client.  Both used a dedicated 10G network.  We basically overran
 our 6030 filer.  We could have jumped through a lot of hoops and split
 backups over multiple weekends, etc, but we decided it wasn't worth
 it.  I haven't tried Flash Backup for a while but it didn't really buy
 us much on what we tried to do with it.  It may work better now.

 We backup everything (~200 TB) to NearStores in another building.  We
 use SnapVault instead of SnapMirror.  We can still revert our
 destination volumes to primary r/w file systems if we need to.  We
 don't have the requirement to send tapes off site.  If you do, you
 could still make

Re: [Veritas-bu] NDMP and Millions of Files

2010-06-17 Thread Ambrose, Monte
You could also use NetBackups Snap Mirror to tape.  You would need a NetApp 
snap mirror license.

The Pros
It uses snapmirror and sends the data off to tape.  It is a RAW volume backup 
and is extremely fast - 10X in many cases.
It can be fully configured in NetBackup

The Cons
It backs up the entire volume - so if you have a 1TB volume and only 400GB are 
used it will backup 1tb
You cant restore a single file or dir.  You have to restore the entire thing.
You cant mix with incremental backups.  

Monte

-Original Message-
From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jonathan Dyck
Sent: Tuesday, June 15, 2010 7:21 AM
To: Jeff Cleverley; rusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

We do our NetApp backups (old GF940 metro cluster) with a combination of 
snapshots (which are available at both sites) and NDMP here. Our NDMP obviously 
holds all the long retention data.

The data size isn't huge compared to some (6.5TB,  ~40M files), but we've had 
to resort to multiple policies and multi-streaming to back it up in a 
reasonable amount of time (less than 60 hours for a full on the wknd).  The way 
it works is:

PolicyA (vol1): explicitly lists 37 different paths for the backup selection 
list,  we've empirically determined these are the smaller folders
PolicyB (vol1-long): explicitly lists 16 paths for the backup selection list,  
we've determined these are the large folders
PolicyC (vol1-Catch-missed-directory): we've mounted the root of vol1 on a 
Linux host,  and we back it up via NFS, excluding the 37+16 paths defined 
above.  If this policy's full backup every gets too large (over 10GB or so),  
we review the contents and add new paths to PolicyA or PolicyB as necessary.  
This is necessary because you can't specify wildcards on NDMP backups 
(discussed in this forum several times I believe).

We repeat the above process for vol2.

The above backup data sits on a deduped VTL for 2 months, and then the data 
that is held longer than that is duplicated to tape for long-term storage and 
expired off the VTL.  As we run 5 streams concurrently, the throughput is 
decent, but we peg out the CPU on the NetApp frequently during the backup 
window,  which is a concern.

HTH...
Jon



-Original Message-
From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jeff Cleverley
Sent: June 14, 2010 5:57 PM
To: rusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

Rusty,

If you have a way to use Snapvault to backup to another location, I
would use it.  We have a number of file systems like what you have.  I
tried NDMP over TCP and NFS backups using dedicated snapshots mounted
on a client.  Both used a dedicated 10G network.  We basically overran
our 6030 filer.  We could have jumped through a lot of hoops and split
backups over multiple weekends, etc, but we decided it wasn't worth
it.  I haven't tried Flash Backup for a while but it didn't really buy
us much on what we tried to do with it.  It may work better now.

We backup everything (~200 TB) to NearStores in another building.  We
use SnapVault instead of SnapMirror.  We can still revert our
destination volumes to primary r/w file systems if we need to.  We
don't have the requirement to send tapes off site.  If you do, you
could still make the tapes from your secondary filer.

Using NFS to tape will give you check points and you can run multiple
streams to each tape drive.  If you are not hitting the filer
throughput limits this may work for you.

Jeff

On Mon, Jun 14, 2010 at 3:24 PM,  rusty.ma...@sungard.com wrote:

 We have a NetApp filer that has a few TB of data made up largely of millions
 of small files (about 30 million or so) and we are using several NDMP
 policies to back up this data. The two main problems are length of time it
 takes to backup (we usually have 2-3 backups running all day every day) and
 when there is a maintenance or other event in the NBU domain, we have to
 kill the job, resulting in having to start all over (no checkpoints).

 For those of you who have faced a similar situation, how are you backing up
 this data?

 Current thoughts are moving away from NDMP and going with just snapshots and
 then getting the snap offsite either by backing it up or replicating it.
 We've also thought about backing it up via NFS, but that will probably be
 slower, though we would get checkpoints.

 I appreciate any other suggestions anyone has.

 Rusty Major, MCSE, BCFP, VCS ▪ Sr. Storage Engineer ▪ SunGard Availability
 Services ▪ 757 N. Eldridge Suite 200, Houston TX 77079 ▪ 281-584-4693
 Keeping People and Information Connected® ▪ http://availability.sungard.com/
 P Think before you print
 CONFIDENTIALITY:  This e-mail (including any attachments) may contain
 confidential

Re: [Veritas-bu] NDMP and Millions of Files

2010-06-17 Thread ccosta . ccc
Good news is that with 7.01 coming out in a month or so will allow NBU to 
multi-stream NDMP data to a single tape drive.

This will/may alleviate some of the performance issues many of you experience 
each day. However I am not sure of any limitations of this feature patch may 
have.

Chris

Sent from my Verizon Wireless BlackBerry

-Original Message-
From: Ambrose, Monte mambr...@qualcomm.com
Date: Thu, 17 Jun 2010 14:38:41 
To: Jonathan Dyckjd...@bank-banque-canada.ca; Jeff 
Cleverleyjeff.clever...@avagotech.com; 
rusty.ma...@sungard.comrusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.eduveritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

You could also use NetBackups Snap Mirror to tape.  You would need a NetApp 
snap mirror license.

The Pros
It uses snapmirror and sends the data off to tape.  It is a RAW volume backup 
and is extremely fast - 10X in many cases.
It can be fully configured in NetBackup

The Cons
It backs up the entire volume - so if you have a 1TB volume and only 400GB are 
used it will backup 1tb
You cant restore a single file or dir.  You have to restore the entire thing.
You cant mix with incremental backups.  

Monte

-Original Message-
From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jonathan Dyck
Sent: Tuesday, June 15, 2010 7:21 AM
To: Jeff Cleverley; rusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

We do our NetApp backups (old GF940 metro cluster) with a combination of 
snapshots (which are available at both sites) and NDMP here. Our NDMP obviously 
holds all the long retention data.

The data size isn't huge compared to some (6.5TB,  ~40M files), but we've had 
to resort to multiple policies and multi-streaming to back it up in a 
reasonable amount of time (less than 60 hours for a full on the wknd).  The way 
it works is:

PolicyA (vol1): explicitly lists 37 different paths for the backup selection 
list,  we've empirically determined these are the smaller folders
PolicyB (vol1-long): explicitly lists 16 paths for the backup selection list,  
we've determined these are the large folders
PolicyC (vol1-Catch-missed-directory): we've mounted the root of vol1 on a 
Linux host,  and we back it up via NFS, excluding the 37+16 paths defined 
above.  If this policy's full backup every gets too large (over 10GB or so),  
we review the contents and add new paths to PolicyA or PolicyB as necessary.  
This is necessary because you can't specify wildcards on NDMP backups 
(discussed in this forum several times I believe).

We repeat the above process for vol2.

The above backup data sits on a deduped VTL for 2 months, and then the data 
that is held longer than that is duplicated to tape for long-term storage and 
expired off the VTL.  As we run 5 streams concurrently, the throughput is 
decent, but we peg out the CPU on the NetApp frequently during the backup 
window,  which is a concern.

HTH...
Jon



-Original Message-
From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jeff Cleverley
Sent: June 14, 2010 5:57 PM
To: rusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

Rusty,

If you have a way to use Snapvault to backup to another location, I
would use it.  We have a number of file systems like what you have.  I
tried NDMP over TCP and NFS backups using dedicated snapshots mounted
on a client.  Both used a dedicated 10G network.  We basically overran
our 6030 filer.  We could have jumped through a lot of hoops and split
backups over multiple weekends, etc, but we decided it wasn't worth
it.  I haven't tried Flash Backup for a while but it didn't really buy
us much on what we tried to do with it.  It may work better now.

We backup everything (~200 TB) to NearStores in another building.  We
use SnapVault instead of SnapMirror.  We can still revert our
destination volumes to primary r/w file systems if we need to.  We
don't have the requirement to send tapes off site.  If you do, you
could still make the tapes from your secondary filer.

Using NFS to tape will give you check points and you can run multiple
streams to each tape drive.  If you are not hitting the filer
throughput limits this may work for you.

Jeff

On Mon, Jun 14, 2010 at 3:24 PM,  rusty.ma...@sungard.com wrote:

 We have a NetApp filer that has a few TB of data made up largely of millions
 of small files (about 30 million or so) and we are using several NDMP
 policies to back up this data. The two main problems are length of time it
 takes to backup (we usually have 2-3 backups running all day every day) and
 when there is a maintenance or other event in the NBU domain, we have to
 kill the job, resulting in having to start all over (no checkpoints).

 For those of you who have faced a similar

Re: [Veritas-bu] NDMP and Millions of Files

2010-06-15 Thread Jonathan Dyck
We do our NetApp backups (old GF940 metro cluster) with a combination of 
snapshots (which are available at both sites) and NDMP here. Our NDMP obviously 
holds all the long retention data.

The data size isn't huge compared to some (6.5TB,  ~40M files), but we've had 
to resort to multiple policies and multi-streaming to back it up in a 
reasonable amount of time (less than 60 hours for a full on the wknd).  The way 
it works is:

PolicyA (vol1): explicitly lists 37 different paths for the backup selection 
list,  we've empirically determined these are the smaller folders
PolicyB (vol1-long): explicitly lists 16 paths for the backup selection list,  
we've determined these are the large folders
PolicyC (vol1-Catch-missed-directory): we've mounted the root of vol1 on a 
Linux host,  and we back it up via NFS, excluding the 37+16 paths defined 
above.  If this policy's full backup every gets too large (over 10GB or so),  
we review the contents and add new paths to PolicyA or PolicyB as necessary.  
This is necessary because you can't specify wildcards on NDMP backups 
(discussed in this forum several times I believe).

We repeat the above process for vol2.

The above backup data sits on a deduped VTL for 2 months, and then the data 
that is held longer than that is duplicated to tape for long-term storage and 
expired off the VTL.  As we run 5 streams concurrently, the throughput is 
decent, but we peg out the CPU on the NetApp frequently during the backup 
window,  which is a concern.

HTH...
Jon



-Original Message-
From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of Jeff Cleverley
Sent: June 14, 2010 5:57 PM
To: rusty.ma...@sungard.com
Cc: veritas-bu@mailman.eng.auburn.edu
Subject: Re: [Veritas-bu] NDMP and Millions of Files

Rusty,

If you have a way to use Snapvault to backup to another location, I
would use it.  We have a number of file systems like what you have.  I
tried NDMP over TCP and NFS backups using dedicated snapshots mounted
on a client.  Both used a dedicated 10G network.  We basically overran
our 6030 filer.  We could have jumped through a lot of hoops and split
backups over multiple weekends, etc, but we decided it wasn't worth
it.  I haven't tried Flash Backup for a while but it didn't really buy
us much on what we tried to do with it.  It may work better now.

We backup everything (~200 TB) to NearStores in another building.  We
use SnapVault instead of SnapMirror.  We can still revert our
destination volumes to primary r/w file systems if we need to.  We
don't have the requirement to send tapes off site.  If you do, you
could still make the tapes from your secondary filer.

Using NFS to tape will give you check points and you can run multiple
streams to each tape drive.  If you are not hitting the filer
throughput limits this may work for you.

Jeff

On Mon, Jun 14, 2010 at 3:24 PM,  rusty.ma...@sungard.com wrote:

 We have a NetApp filer that has a few TB of data made up largely of millions
 of small files (about 30 million or so) and we are using several NDMP
 policies to back up this data. The two main problems are length of time it
 takes to backup (we usually have 2-3 backups running all day every day) and
 when there is a maintenance or other event in the NBU domain, we have to
 kill the job, resulting in having to start all over (no checkpoints).

 For those of you who have faced a similar situation, how are you backing up
 this data?

 Current thoughts are moving away from NDMP and going with just snapshots and
 then getting the snap offsite either by backing it up or replicating it.
 We've also thought about backing it up via NFS, but that will probably be
 slower, though we would get checkpoints.

 I appreciate any other suggestions anyone has.

 Rusty Major, MCSE, BCFP, VCS ▪ Sr. Storage Engineer ▪ SunGard Availability
 Services ▪ 757 N. Eldridge Suite 200, Houston TX 77079 ▪ 281-584-4693
 Keeping People and Information Connected® ▪ http://availability.sungard.com/
 P Think before you print
 CONFIDENTIALITY:  This e-mail (including any attachments) may contain
 confidential, proprietary and privileged information, and unauthorized
 disclosure or use is prohibited.  If you received this e-mail in error,
 please notify the sender and delete this e-mail from your system.
 ___
 Veritas-bu maillist  -  veritas...@mailman.eng.auburn.edu
 http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu





-- 
Jeff Cleverley
Unix Systems Administrator
4380 Ziegler Road
Fort Collins, Colorado 80525
970-288-4611
___
Veritas-bu maillist  -  Veritas-bu@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu


La version française suit le texte anglais

Re: [Veritas-bu] NDMP and Millions of Files

2010-06-14 Thread Martin, Jonathan
We’ve been testing backup methods for millions of small files on NAS and DAS 
storage for YEARS here. Without exception, the best backup method has been NFS 
mount via a “proxy” client. I don’t manage the NetApp Storage, but we saw an 
increase from 10MB/sec to 17-19MB/sec when switching from NDMP to NFS on a 
FAS2040. I’ve heard there is a FAS 31XX coming to replace the FAS2040 and 
hopefully that gives us better performance still. Using NFS gets you 
checkpoints, but more importantly for us, exclusions. Our filers have something 
like 70-100 million files, but we only backup 45-50 million of them.

 

Off the top of my head, my best test runs per device are:

Apple xRaid DAS (unknown model) – 30MB/sec via 2 streams NFS

Dell MD3000 DAS (RedHat 5) – 30MB/sec via single stream

Sun Unified Storage System 7210 – 25-29MB/sec via 16 streams via NFS

NetApp FAS2040 – 19MB/sec 

Sun StorageTek 5310 – 8MB/sec via single stream NFS

 

We’re currently using the Dell MD3000 / Redhat 5 rig above to test Linux 
Flashbackup, but so far we’re only running 24MB/sec.

 

The most important lesson I’ve learned converting NDMP to NFS is to test 
thoroughly. Every filer is different, and it can take weeks to find a “sweet 
spot” between number or streams and client configuration.

 

Good luck!

 

-Jonathan

 

 

From: veritas-bu-boun...@mailman.eng.auburn.edu 
[mailto:veritas-bu-boun...@mailman.eng.auburn.edu] On Behalf Of 
rusty.ma...@sungard.com
Sent: Monday, June 14, 2010 5:25 PM
To: veritas-bu@mailman.eng.auburn.edu
Subject: [Veritas-bu] NDMP and Millions of Files

 


We have a NetApp filer that has a few TB of data made up largely of millions of 
small files (about 30 million or so) and we are using several NDMP policies to 
back up this data. The two main problems are length of time it takes to backup 
(we usually have 2-3 backups running all day every day) and when there is a 
maintenance or other event in the NBU domain, we have to kill the job, 
resulting in having to start all over (no checkpoints). 

For those of you who have faced a similar situation, how are you backing up 
this data? 

Current thoughts are moving away from NDMP and going with just snapshots and 
then getting the snap offsite either by backing it up or replicating it. We've 
also thought about backing it up via NFS, but that will probably be slower, 
though we would get checkpoints. 

I appreciate any other suggestions anyone has. 

Rusty Major, MCSE, BCFP, VCS ▪ Sr. Storage Engineer ▪ SunGard Availability 
Services ▪ 757 N. Eldridge Suite 200, Houston TX 77079 ▪ 281-584-4693 
Keeping People and Information Connected® ▪ http://availability.sungard.com/ 
http://availability.sungard.com/  
P Think before you print 
CONFIDENTIALITY:  This e-mail (including any attachments) may contain 
confidential, proprietary and privileged information, and unauthorized 
disclosure or use is prohibited.  If you received this e-mail in error, please 
notify the sender and delete this e-mail from your system. 

___
Veritas-bu maillist  -  Veritas-bu@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu


Re: [Veritas-bu] NDMP and Millions of Files

2010-06-14 Thread Jeff Cleverley
Rusty,

If you have a way to use Snapvault to backup to another location, I
would use it.  We have a number of file systems like what you have.  I
tried NDMP over TCP and NFS backups using dedicated snapshots mounted
on a client.  Both used a dedicated 10G network.  We basically overran
our 6030 filer.  We could have jumped through a lot of hoops and split
backups over multiple weekends, etc, but we decided it wasn't worth
it.  I haven't tried Flash Backup for a while but it didn't really buy
us much on what we tried to do with it.  It may work better now.

We backup everything (~200 TB) to NearStores in another building.  We
use SnapVault instead of SnapMirror.  We can still revert our
destination volumes to primary r/w file systems if we need to.  We
don't have the requirement to send tapes off site.  If you do, you
could still make the tapes from your secondary filer.

Using NFS to tape will give you check points and you can run multiple
streams to each tape drive.  If you are not hitting the filer
throughput limits this may work for you.

Jeff

On Mon, Jun 14, 2010 at 3:24 PM,  rusty.ma...@sungard.com wrote:

 We have a NetApp filer that has a few TB of data made up largely of millions
 of small files (about 30 million or so) and we are using several NDMP
 policies to back up this data. The two main problems are length of time it
 takes to backup (we usually have 2-3 backups running all day every day) and
 when there is a maintenance or other event in the NBU domain, we have to
 kill the job, resulting in having to start all over (no checkpoints).

 For those of you who have faced a similar situation, how are you backing up
 this data?

 Current thoughts are moving away from NDMP and going with just snapshots and
 then getting the snap offsite either by backing it up or replicating it.
 We've also thought about backing it up via NFS, but that will probably be
 slower, though we would get checkpoints.

 I appreciate any other suggestions anyone has.

 Rusty Major, MCSE, BCFP, VCS ▪ Sr. Storage Engineer ▪ SunGard Availability
 Services ▪ 757 N. Eldridge Suite 200, Houston TX 77079 ▪ 281-584-4693
 Keeping People and Information Connected® ▪ http://availability.sungard.com/
 P Think before you print
 CONFIDENTIALITY:  This e-mail (including any attachments) may contain
 confidential, proprietary and privileged information, and unauthorized
 disclosure or use is prohibited.  If you received this e-mail in error,
 please notify the sender and delete this e-mail from your system.
 ___
 Veritas-bu maillist  -  veritas...@mailman.eng.auburn.edu
 http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu





-- 
Jeff Cleverley
Unix Systems Administrator
4380 Ziegler Road
Fort Collins, Colorado 80525
970-288-4611
___
Veritas-bu maillist  -  Veritas-bu@mailman.eng.auburn.edu
http://mailman.eng.auburn.edu/mailman/listinfo/veritas-bu