Re: Rsync 4TB datafiles...?

2006-05-18 Thread John Van Essen
On Wed, 17 May 2006, lsk [EMAIL PROTECTED] wrote:
 
 So this shows that RSYNC performs faster when there is no datafile is
 present. For Oracle datafiles the diff algorithm doesn't work even with the
 patch. Literally it should have taken less than 12 min to just compare and
 update the header changes but it took more than 12 min when the file is
 already present at destination(target).

Are you using the --inplace option?

If the file doesn't exist, the entire file is sent, yes, but there's
only a serial write operation needed at the destination.

If the file *does* exist, rsync must read the existing file in its
entirety while *also* writing out the new file.  The dance that the
disk R/W head might be doing in this case might slow things down, too.

 I am not going to give up I am going to try again and see...

OK.  :)

John

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-05-17 Thread lsk

TEST RESULTS after applying the patch Dynamic_hash.diff with new rsync
version 2.6.8.
--

The patch was installed on the source server and I tested with a 2 GB file
transfer with following options and these are the timings...

1. No DESTINATION File - 2 GB - 12 min
2. With DESTINATION File nothing changed - 2 GB - 21 min
3. With DESTINATION File header information changed - 17 min

The first 2 were tested with rysnc -zv and the last one with rsync -zv
--no-whole-file

So this shows that RSYNC performs faster when there is no datafile is
present. For Oracle datafiles the diff algorithm doesn't work even with the
patch. Literally it should have taken less than 12 min to just compare and
update the header changes but it took more than 12 min when the file is
already present at destination(target).

I am not going to give up I am going to try again and see...

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a4429359
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-05-02 Thread lsk

Also I use the rsync version rsync  version 2.6.5  protocol version 29 does
this version include this patch dynamic_hash.diff or do we need to
install it seperately.
  

Sorry. You will need to get the 2.6.7 sources, and then apply the patch
yourself and compile rsync.

Please do report back here your results. This patch is a result of a lot
of theoretical work, but we never got any actual feedback on it.

   Shachar
-- 

***Hello Shachar...is 2.6.7 is the latest version of rsync. I could see
in the http download site it says rsync-2.6.8.tar.gz. Should I get this
version 2.6.8 + the patch dynamic_hash.diff. 

Also I am planning to install in only the sending machine...and first try
out.

Thanks for your feedback.
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a4180739
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-05-02 Thread Shachar Shemesh
lsk wrote:

Hello Shachar...is 2.6.7 is the latest version of rsync. I could see
in the http download site it says rsync-2.6.8.tar.gz. Should I get this
version 2.6.8 + the patch dynamic_hash.diff. 
  

Yes. In the over a month that passed since the email I sent a new
version of rsync was released :-)

Dynamic_hash.diff is available in that one too.

Also I am planning to install in only the sending machine...and first try
out.
  

Should work.

Thanks for your feedback.
lsk.

   Shachar

-- 
Shachar Shemesh
Lingnu Open Source Consulting ltd.
Have you backed up today's work? http://www.lingnu.com/backup.html

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-05-02 Thread lsk


lsk wrote:

Hello Shachar...is 2.6.7 is the latest version of rsync. I could see
in the http download site it says rsync-2.6.8.tar.gz. Should I get this
version 2.6.8 + the patch dynamic_hash.diff. 
  

Yes. In the over a month that passed since the email I sent a new
version of rsync was released :-)

Dynamic_hash.diff is available in that one too.

//***Shachar is Dynamic_hash.diff in built in 2.6.8 code version or still
I need to apply that as a seperate patch after I install rsync  2.6.8 ?

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a4198608
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-05-02 Thread Matt McCutchen
On Tue, 2006-05-02 at 13:06 -0700, lsk wrote:
 //***Shachar is Dynamic_hash.diff in built in 2.6.8 code version or still
 I need to apply that as a seperate patch after I install rsync  2.6.8 ?

You need to apply it as a separate patch _before you build_ rsync 2.6.8.
Apply the patch by running
patch -p1 patches/dynamic_hash.diff
in the source tree.  Then configure, make, and install.

Matt

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-23 Thread Paul Slootman
On Wed 22 Mar 2006, Linus Hicks wrote:
 Paul Slootman wrote:
 
 I'd recommend doing --inplace, as chances are that data won't move
 within a file with oracle data files (so it's not useful to try to find
 moved data), and copying the 4TB to temp. files every time could become
 a big timewaster. Also the -t option could be handy, not all files
 change all the time IIRC.
 
 The above remark about not being useful to try to find moved data 
 provoked an idea. But my understanding of --inplace is apparently different 
 from yours. I thought --inplace only meant that the destination file would 
 be directly overwritten, not that it would turn off any of the 
 optimizations for finding moved data.

I go on what's in the manpage:

 --inplace
This causes rsync not to create a new copy of the file and then move  it
into  place.   Instead  rsync  will overwrite the existing file, meaning
that the rsync algorithm can't accomplish the  full  amount  of  network
reduction  it  might  be able to otherwise (since it does not yet try to
sort data matches).  One exception to this is if you combine the  option
with --backup, since rsync is smart enough to use the backup file as the
basis file for the transfer.


Paul Slootman
-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-22 Thread lsk


lsk wrote:

But I have tried various options including --inplace,--no-whole-file etc.,
for last few weeks but all the results show me removing the destination
server oracle datafiles and after that doing an rsync -vz from source is
faster than copying(rsyncing) over the old files that are present in
destination.
  

Please do try applying the patch in patches/dynamic_hash.diff to both
sides (well, it's probably only necessary for the sending machine, but
no matter) and making this check again. This patch is meant to address
precisely your predicament.

  Shachar
-- 
/// lsk  Hi which issue does this patch address ? it is for -- inplace 
or for ,--no-whole-file transfer of oracle datafiles. 

Where can I get this patch ?

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a3534475
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-22 Thread lsk

Also I use the rsync version rsync  version 2.6.5  protocol version 29 does
this version include this patch dynamic_hash.diff or do we need to
install it seperately.

At destination server I use rsync  version 2.6.6  protocol version 29
anyhow u said that doesn't matter.

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a3534654
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-22 Thread Linus Hicks

Paul Slootman wrote:

On Tue 21 Mar 2006, lsk wrote:


I don't know how it would work if we do rsync with the files--from option ?


I'm not sure how rsync behaves when confronted with a network problem
during a session, so I won't give an answer to that.
However, doing individual files sounds reasonable, so make it a loop:

 dbf-list while read filename; do rsync -vz $filename destser:$filename done


Also rsync gurus would you suggest which is the fastest way to trasfer this
4 TB data ? Any suggestions...would be of great help.


I'd recommend doing --inplace, as chances are that data won't move
within a file with oracle data files (so it's not useful to try to find
moved data), and copying the 4TB to temp. files every time could become
a big timewaster. Also the -t option could be handy, not all files
change all the time IIRC.


The above remark about not being useful to try to find moved data provoked an 
idea. But my understanding of --inplace is apparently different from yours. I 
thought --inplace only meant that the destination file would be directly 
overwritten, not that it would turn off any of the optimizations for finding 
moved data.


It would be useful (I think) on a fast network to be able to turn off those 
optimizations, and only compare blocks located at the same offset in source and 
destination. If that is not how --inplace works, I wonder if that would be a 
performance win.


Linus
--
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-22 Thread Shachar Shemesh
lsk wrote:

Also I use the rsync version rsync  version 2.6.5  protocol version 29 does
this version include this patch dynamic_hash.diff or do we need to
install it seperately.
  

Sorry. You will need to get the 2.6.7 sources, and then apply the patch
yourself and compile rsync.

Please do report back here your results. This patch is a result of a lot
of theoretical work, but we never got any actual feedback on it.

   Shachar
-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Rsync 4TB datafiles...?

2006-03-21 Thread lsk

I need to rsync 4 TB datafiles to remote server and clone to a new oracle
database..I have about 40 drives that contains this 4 TB data. I would like
to do rsync from a directory level by using --files-from=FILE option. But
the problem is what will happen if the network connection fails the whole
rsync will fail right.

rsync -a srchost:/ / --files-from=dbf-list 

and dbf-list would contain this: 

p02/oradata/OSID/redo01.log 
p03/oradata/OSID/redo02.log 
p04/oradata/OSID/redo03.log 
p01/oradata/OSID/system01.dbf 
p04/oradata/OSID/undotbs01.dbf 
p03/oradata/OSID/sysaux01.dbf 
p03/oradata/OSID/users01.dbf 
p03/oradata/OSID/example01.dbf 
p03/oradata/OSID/data01.dbf 
p02/oradata/OSID/index01.dbf 
p01/oradata/OSID/control01.ctl 
p02/oradata/OSID/control02.ctl 
p03/oradata/OSID/control03.ctl 
s01/oracle/product/10.2.0/dbs/initOSID.ora 


One advantage that I have if do rsync file by file is if the network
connection fails for a quite sometime and comes back alive the rsync will
continue from the next file in the list while it will fail on the other
files when connection is lost.

rsync -vz /p02/oradata/OSID/redo01.log 
destser:/d02/oradata/OSID/redo01.log
rsync -vz /p03/oradata/OSID/redo02.log 
destser:/d03/oradata/OSID/redo02.log   
rsync -vz /p04/oradata/OSID/redo03.log 
destser:/d04/oradata/OSID/redo03.log   
rsync -vz /p01/oradata/OSID/system01.dbf   
destser:/d01/oradata/OSID/system01.dbf 
rsync -vz /p04/oradata/OSID/undotbs01.dbf  
destser:/d04/oradata/OSID/undotbs01.dbf
rsync -vz /p03/oradata/OSID/sysaux01.dbf   
destser:/d03/oradata/OSID/sysaux01.dbf 
rsync -vz /p03/oradata/OSID/users01.dbf
destser:/d03/oradata/OSID/users01.dbf  
rsync -vz /p03/oradata/OSID/example01.dbf  
destser:/d03/oradata/OSID/example01.dbf
rsync -vz /p03/oradata/OSID/data01.dbf 
destser:/d03/oradata/OSID/data01.dbf   
rsync -vz /p02/oradata/OSID/index01.dbf
destser:/d02/oradata/OSID/index01.dbf  
rsync -vz /p01/oradata/OSID/control01.ctl  
destser:/d01/oradata/OSID/control01.ctl
rsync -vz /p02/oradata/OSID/control02.ctl  
destser:/d02/oradata/OSID/control02.ctl
rsync -vz /p03/oradata/OSID/control03.ctl  
destser:/d03/oradata/OSID/control03.ctl
rsync -vz /s01/oracle/product/10.2.0/dbs/initOSID.ora  
destser:/d01/oracle/droduct/10.2.0/dbs/initOSID.ora

I don't know how it would work if we do rsync with the files--from option ?
Also rsync gurus would you suggest which is the fastest way to trasfer this
4 TB data ? Any suggestions...would be of great help.

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a3516542
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-21 Thread Paul Slootman
On Tue 21 Mar 2006, lsk wrote:

 I don't know how it would work if we do rsync with the files--from option ?

I'm not sure how rsync behaves when confronted with a network problem
during a session, so I won't give an answer to that.
However, doing individual files sounds reasonable, so make it a loop:

 dbf-list while read filename; do rsync -vz $filename destser:$filename done

 Also rsync gurus would you suggest which is the fastest way to trasfer this
 4 TB data ? Any suggestions...would be of great help.

I'd recommend doing --inplace, as chances are that data won't move
within a file with oracle data files (so it's not useful to try to find
moved data), and copying the 4TB to temp. files every time could become
a big timewaster. Also the -t option could be handy, not all files
change all the time IIRC.

As always, get the latest version.


Paul Slootman
-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-21 Thread lsk

 Also rsync gurus would you suggest which is the fastest way to trasfer this 
 4 TB data ? Any suggestions...would be of great help. 

I'd recommend doing --inplace, as chances are that data won't move 
within a file with oracle data files (so it's not useful to try to find 
moved data), and copying the 4TB to temp. files every time could become 
a big timewaster. Also the -t option could be handy, not all files 
change all the time IIRC. 

***lsk 

I can't use -t because the oracle datafiles header information are constanly
updated so it will definetely change.

But I have tried various options including --inplace,--no-whole-file etc.,
for last few weeks but all the results show me removing the destination
server oracle datafiles and after that doing an rsync -vz from source is
faster than copying(rsyncing) over the old files that are present in
destination.

I put list of rsync file commands in a batch file and spawn them so that
they run in background.

Thanks,
lsk.
--
View this message in context: 
http://www.nabble.com/Rsync-4TB-datafiles...--t1318624.html#a3518597
Sent from the Samba - rsync forum at Nabble.com.

-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html


Re: Rsync 4TB datafiles...?

2006-03-21 Thread Shachar Shemesh
lsk wrote:

But I have tried various options including --inplace,--no-whole-file etc.,
for last few weeks but all the results show me removing the destination
server oracle datafiles and after that doing an rsync -vz from source is
faster than copying(rsyncing) over the old files that are present in
destination.
  

Please do try applying the patch in patches/dynamic_hash.diff to both
sides (well, it's probably only necessary for the sending machine, but
no matter) and making this check again. This patch is meant to address
precisely your predicament.

  Shachar
-- 
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html