I have found that a tar pipe is much faster than rsync for this sort of thing. 
The fastest of these is ‘star’ (schily tar). On average it is about 2x-5x 
faster than rsync for doing this. After one pass with this, you can use rsync 
for a subsequent or last pass synch.

e.g.
$ cd /export/gpfs1/foo
$ star –c H=xtar | (cd /export/gpfs2/foo; star –xp)

This also will not preserve filesets and quotas, though. You should be able to 
automate that with a little bit of awk, perl, or whatnot.


From: [email protected] 
[mailto:[email protected]] On Behalf Of Damir Krstic
Sent: Friday, January 29, 2016 2:32 PM
To: gpfsug main discussion list
Subject: [gpfsug-discuss] migrating data from GPFS3.5 to ESS appliance (GPFS4.1)

We have recently purchased ESS appliance from IBM (GL6) with 1.5PT of storage. 
We are in planning stages of implementation. We would like to migrate date from 
our existing GPFS installation (around 300TB) to new solution.

We were planning of adding ESS to our existing GPFS cluster and adding its 
disks and then deleting our old disks and having the data migrated this way. 
However, our existing block size on our projects filesystem is 1M and in order 
to extract as much performance out of ESS we would like its filesystem created 
with larger block size. Besides rsync do you have any suggestions of how to do 
this without downtime and in fastest way possible?

I have looked at AFM but it does not seem to migrate quotas and filesets so 
that may not be an optimal solution.
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to