Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread David Rees
On 3/26/07, Evren Yurtesen [EMAIL PROTECTED] wrote:
 Lets hope this doesnt wrap around... as you can see load is in 0.1-0.01
 range.

  1 usersLoad  0.12  0.05  0.01  Mar 27 07:30

 Mem:KBREALVIRTUAL VN PAGER  SWAP PAGER
  Tot   Share  TotShareFree in  out in  out
 Act   260203592   144912 6868   12384 count
 All  2497845456  232789611800 pages

It wrapped pretty badly, but let me see if I'm interpreting this right
(I'm no BSD expert, either):

1. Your server has ~250MB of memory.
2. Load average during backups is only 0.1-0.01? Does BSD calculate
load average differently than Linux? Linux calculates load average by
looking at the number of runnable tasks - this means if you have a
single process waiting on disk IO you will have a load average of 1.
If BSD calculates the load average the same way, then that means your
server is not waiting on disk, but waiting for the clients.

What's the load like on the clients you are backing up?

-Dave

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread John Pettitt
David Rees wrote:
 On 3/26/07, Evren Yurtesen [EMAIL PROTECTED] wrote:
   
 Lets hope this doesnt wrap around... as you can see load is in 0.1-0.01
 range.

  1 usersLoad  0.12  0.05  0.01  Mar 27 07:30

 Mem:KBREALVIRTUAL VN PAGER  SWAP PAGER
  Tot   Share  TotShareFree in  out in  out
 Act   260203592   144912 6868   12384 count
 All  2497845456  232789611800 pages
 

 It wrapped pretty badly, but let me see if I'm interpreting this right
 (I'm no BSD expert, either):

 1. Your server has ~250MB of memory.
 2. Load average during backups is only 0.1-0.01? Does BSD calculate
 load average differently than Linux? Linux calculates load average by
 looking at the number of runnable tasks - this means if you have a
 single process waiting on disk IO you will have a load average of 1.
 If BSD calculates the load average the same way, then that means your
 server is not waiting on disk, but waiting for the clients.

 What's the load like on the clients you are backing up?
   
Under FreeBSD if a task is blocked it doesn't contribute to load no 
matter what it's waiting for - so disk bound tasks don't add to load 
average.

John



-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
brien dieterle wrote:
 Jason Hughes wrote:
 Evren Yurtesen wrote:
   
 Jason Hughes wrote:
 
 That drive should be more than adequate.  Mine is a 5400rpm 2mb 
 buffer clunker.  Works fine.
 Are you running anything else on the backup server, besides 
 BackupPC?  What OS?  What filesystem?  How many files total?
   
 FreeBSD, UFS2+softupdates, noatime.

 There are 4 hosts that have been backed up, for a total of:

 * 16 full backups of total size 72.16GB (prior to pooling and 
 compression),
 * 24 incr backups of total size 13.45GB (prior to pooling and 
 compression).

 # Pool is 17.08GB comprising 760528 files and 4369 directories (as of 
 3/27 05:54),
 # Pool hashing gives 38 repeated files with longest chain 6,
 # Nightly cleanup removed 10725 files of size 0.40GB (around 3/27 05:54),
 # Pool file system was recently at 10% (3/27 07:16), today's max is 
 10% (3/27 01:00) and yesterday's max was 10%.

  Host   User   #Full   Full Age (days)   Full Size 
 (GB)   Speed (MB/s)   #Incr   Incr Age (days)   Last 
 Backup (days)   State   Last attempt
 host1 4 5.4 3.88 0.22 6 0.4 
 0.4 idle idle
 host2 4 5.4 2.10 0.06 6 0.4 
 0.4 idle idle
 host3 4 5.4 7.57 0.14 6 0.4 
 0.4 idle idle
 host4 4 5.4 5.56 0.10 6 0.4 
 0.4 idle idle

 
 Hmm.  This is a tiny backup setup, even smaller than mine.  However, it 
 appears that the average size of your file is only 22KB, which is quite 
 small.  For comparison sake, this is from my own server:
 Pool is 172.91GB comprising 217311 files and 4369 directories (as of 
 3/26 01:08),

 The fact that you have tons of little files will probably give 
 significantly higher overhead when doing file-oriented work, simply 
 because the inodes must be fetched for each file before seeking to the 
 file itself.  If we assume no files are shared between hosts (very 
 conservative), and you have an 8ms access time, you will have 190132 
 files per host and two seeks per file, neglecting actual i/o time, gives 
 you 50 minutes.  Just to seek them all.  If you have a high degree of 
 sharing, it can be up to 4x worse.  Realize, the same number of seeks 
 must be made on the server as well as the client.

 Are you sure you need to be backing up everything that you're putting 
 across the network?  Maybe excluding some useless directories, maybe 
 temp files or logs that haven't been cleaned up?  Perhaps you can 
 archive big chunks of it with a cron job?

 I'd start looking for ways to cut down the number of files, because the 
 overhead of per-file accesses are probably eating you alive.  I'm also 
 no expert on UFS2 or FreeBSD, so it may be worthwhile to research its 
 behavior with hard links and small files.

 JH

   
 For what it's worth, I have a server that backs up 8.6 million files  
 averaging 10k in size from one host.  It takes a full 10 hours for a 
 full backup via tar over NFS ( 2.40MB/s for 87GB). CPU usage is low, 
 around 10-20%, however IOwait is a pretty steady 25%.
 
 Server info:
 HP DL380 G4
 debian sarge
 dual processor 3.2ghz xeon
 2GB Ram
 5x10k rpm scsi disks, raid5
 128MB battery backed cache (50/50 r/w)
 ext3 filesystems
 
 brien

You are distributing the reads and writes on 5 disks here. Dont you 
think that 2.40mb/s is a little bit slow compared to the horsepower you 
have?

If you scale it down to my system, I think 1/10 performance is normal...

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
Craig Barratt wrote:
 Evren writes:
 
   HostUser#Full   Full Age (days) Full Size (GB) 
  Speed 
 (MB/s)#Incr   Incr Age (days) Last Backup (days) 
  State   
 Last attempt
 host14   5.4 3.880.226   0.4 
 0.4 idleidle
 host24   5.4 2.100.066   0.4 
 0.4 idleidle
 host34   5.4 7.570.146   0.4 
 0.4 idleidle
 host44   5.4 5.560.106   0.4 
 0.4 idleidle
 
 These xfer rates are certainly very low.
 
 I don't think I caught which XferMethod you are using.  Also,
 what is your BackupPC version?  If you are using 3.0, what is
 the value of $Conf{IncrLevels}?
 
 Craig

I use rsync with 3.0.0 but it was the same speed with 2.x.x versions.

$Conf{IncrLevels} = [1];

I wonder, since I have:

$Conf{IncrKeepCnt} = 6;

Wouldnt it make more sense to use this?:

$Conf{IncrLevels}  = [1, 2, 3, 4, 5, 6];

or does this make things even more slower?

I have concluded that to get a decent performance from backuppc one must 
use RAID and put a lot of ram to the machine because BackupPC is using 
the disk intensively!

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Evren Yurtesen wrote:

 Server info:
 HP DL380 G4
 debian sarge
 dual processor 3.2ghz xeon
 2GB Ram
 5x10k rpm scsi disks, raid5
 128MB battery backed cache (50/50 r/w)
 ext3 filesystems

 brien
 
 You are distributing the reads and writes on 5 disks here. Dont you 
 think that 2.40mb/s is a little bit slow compared to the horsepower you 
 have?

Raid5 doesn't distribute disk activity - it puts the drives in lockstep 
and is slower than a single drive, especially on small writes where it 
has to do extra reads to re-compute parity on the existing data.

 If you scale it down to my system, I think 1/10 performance is normal...

Its the RAM and disk cache that makes this system work.

-- 
   Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
Les Mikesell wrote:

 That doesn't sound difficult at all.  I suspect your real problem is 
 that you are running a *bsd UFS filesystem with it's default sync 
 metadata handling which is going to wait for the physical disk action to 
 complete on every directory operation.  I think there are other options 
 but I haven't kept up with them.  I gave up on UFS long ago when I 
 needed to make an application that frequently truncated and rewrote a 
 data file work on a machine that crashed frequently.  The sync-metadata 
 'feature' statistically ensured that there was never any data in the 
 file after recovering since the truncation was always forced to disk 
 immediately but the data write was buffered so with a fast cycle the 
 on-disk copy was nearly always empty.
 
 Is anyone else running a *bsd?
 

The problem is the reads, not writes. It takes long time for BackupPC to 
figure out if the file should be backed up or not. Backups take very 
long time even when there are few files which are backed up.

But I have to disagree, again :) When you make writes and filesystem is 
mounted with 'sync' option which is default on FreeBSD then all 
operations are done sync. Your program must have been doing something 
between truncate and the actual write operation which caused file to be 
empty for a period of time.

 From FreeBSD mount manual
- syncAll I/O to the file system should be done synchronously.
- async   All I/O to the file system should be done asynchronously.
This is a dangerous flag to set, and should not be used
unless you are prepared to recreate the file system
should your system crash.

This 'feature' as you call it, exists in linux also

 From Linux mount manual
- sync   All I/O to the file system should be done  synchronously
- async  All I/O to the file system should be done asynchronously.

The only difference is that FreeBSD uses sync by default because they 
claim(there is no point to argue if it is true or not, there are enough 
threads about those if you google :) ) that it is safer while Linux uses 
async. You could easily enable 'async' IO on FreeBSD to have the same 
behaviour.

 Perhaps, but there is a difference if they are moving 10 times or 
 10 times. Where the difference is that the possibility of 
 failure due to mechanical problems increases 1 times.

 No, it doesn't make a lot of difference as long as the drive doesn't 
 overheat.  The head only moves so fast and it doesn't matter if it 
 does it continuously.  However, if your system has sufficient RAM, it 
 will cache and optimize many of the things that might otherwise need 
 an additional seek and access.

 I cant see how you can reach to this conclusion.
 
 Observation... I run hundreds of servers, many of which are 5 or more 
 years old.  The disk failures have had no correlation to the server 
 activity.

Hard drive manufacturers disagree with you. They do consider the 
duty-cycles when calculating the MTBF values.
http://www.digit-life.com/articles/storagereliability/

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
David Rees wrote:
 On 3/26/07, Evren Yurtesen [EMAIL PROTECTED] wrote:
 Lets hope this doesnt wrap around... as you can see load is in 0.1-0.01
 range.

  1 usersLoad  0.12  0.05  0.01  Mar 27 07:30

 Mem:KBREALVIRTUAL VN PAGER  SWAP PAGER
  Tot   Share  TotShareFree in  out in  out
 Act   260203592   144912 6868   12384 count
 All  2497845456  232789611800 pages
 
 It wrapped pretty badly, but let me see if I'm interpreting this right
 (I'm no BSD expert, either):
 
 1. Your server has ~250MB of memory.
 2. Load average during backups is only 0.1-0.01? Does BSD calculate
 load average differently than Linux? Linux calculates load average by
 looking at the number of runnable tasks - this means if you have a
 single process waiting on disk IO you will have a load average of 1.
 If BSD calculates the load average the same way, then that means your
 server is not waiting on disk, but waiting for the clients.
 
 What's the load like on the clients you are backing up?
 
 -Dave

2- The process is idle when it is waiting data from the disk (not in 
active state) So FreeBSD doesnt count it towards the load. I dont see 
why it should actually? Are processes which are waiting data from disk 
use CPU in Linux?

1- You are right, I will increase the memory a little bit. I will return 
back the results when I accomplish that. :) Lets hope it will decrease 
disk reads.

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
Les Mikesell wrote:
 Evren Yurtesen wrote:
 
 Server info:
 HP DL380 G4
 debian sarge
 dual processor 3.2ghz xeon
 2GB Ram
 5x10k rpm scsi disks, raid5
 128MB battery backed cache (50/50 r/w)
 ext3 filesystems

 brien

 You are distributing the reads and writes on 5 disks here. Dont you 
 think that 2.40mb/s is a little bit slow compared to the horsepower 
 you have?
 
 Raid5 doesn't distribute disk activity - it puts the drives in lockstep 
 and is slower than a single drive, especially on small writes where it 
 has to do extra reads to re-compute parity on the existing data.

I am confused, when a write is done the data is distributed in the disks 
depending on the stripe size you are using. When you start reading the 
file, you are reading from 5 different disks. So you get way better 
performance for sure on reads.

Agreed, writes might be slower depending on the raid controller, amount 
of cache it has etc. but my problem here is the reads. In my machine 
BackupPC server is working slow even when it is backing up a very small 
amount of files.

 If you scale it down to my system, I think 1/10 performance is normal...
 
 Its the RAM and disk cache that makes this system work.
 

Agreed, I will add more RAM and report the results.

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Evren Yurtesen wrote:

 The problem is the reads, not writes. It takes long time for BackupPC to 
 figure out if the file should be backed up or not. Backups take very 
 long time even when there are few files which are backed up.

That's only true with rsync and especially rsync fulls where the 
contents are compared as well as the timestamps. But I thought your 
complaint was about the hardlinks and data storage.  The transfer speed 
value doesn't make a lot of sense in the rsync case anyway because you 
often spend a lot of time computing to avoid having to transfer anything 
- and you have the entire directory in RAM while doing it which may 
reduce your disk buffering.  What is wall clock time for a run and is it 
reasonable for having to read through both the client and server copies?

 But I have to disagree, again :) When you make writes and filesystem is 
 mounted with 'sync' option which is default on FreeBSD then all 
 operations are done sync. Your program must have been doing something 
 between truncate and the actual write operation which caused file to be 
 empty for a period of time. 

Yes, it was writing a small amount to the file and the data write was 
always deferred with this kind of problem:
http://www.complang.tuwien.ac.at/anton/sync-metadata-updates.html

 The only difference is that FreeBSD uses sync by default because they 
 claim(there is no point to argue if it is true or not, there are enough 
 threads about those if you google :) ) that it is safer while Linux uses 
 async.

It is safer for the file system if you are concerned that fsck can't fix 
it.  It is not safer for the data of any particular file.  Anyway 
crashes are pretty much a thing of the past and I protect against most 
kinds of failures by raid-mirroring to an external drive that is rotated 
off-site weekly.  I'm running a journaled reiserfs because it is very 
fast at creating and deleting files, but ext3 with the right options 
should also work.

 Perhaps, but there is a difference if they are moving 10 times or 
 10 times. Where the difference is that the possibility of 
 failure due to mechanical problems increases 1 times.

 No, it doesn't make a lot of difference as long as the drive doesn't 
 overheat.  The head only moves so fast and it doesn't matter if it 
 does it continuously.  However, if your system has sufficient RAM, 
 it will cache and optimize many of the things that might otherwise 
 need an additional seek and access.

 I cant see how you can reach to this conclusion.

 Observation... I run hundreds of servers, many of which are 5 or more 
 years old.  The disk failures have had no correlation to the server 
 activity.
 
 Hard drive manufacturers disagree with you. They do consider the 
 duty-cycles when calculating the MTBF values.
 http://www.digit-life.com/articles/storagereliability/

Vendor MTBF values are very unrealistic. The trade rags seem to have 
caught on to this recently: 
http://www.eweek.com/article2/0%2C1895%2C2105551%2C00.asp
If you want to know what the vendor really expects, look at the warranty 
instead.

-- 
   Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Mike Dresser
On Mon, 26 Mar 2007, David Rees wrote:

 No kidding! My backuppc stats are like this:

For my biggest server, my stats are:

There are 36 hosts that have been backed up, for a total of:

503 full backups of total size 4200.30GB (prior to pooling and compression),
247 incr backups of total size 444.52GB (prior to pooling and compression).

Pool is 776.90GB comprising 1900158 files and 4369 directories

Dual Opteron 265(4x1.8ghz cores), 1 GB ram, 3ware 9550SX-16ML w/15 Maxtor 
MaxLine III 300 (7200 rpm, ATA, 8 meg cache), ext3 w/noatime.  I'd change 
it to XFS, but backing up 1.7TB would take weeks/months because of the 
hardlinks.

Runs BackupPC 2.1.2-3 on Debian Sarge.

Backups take about 5-6 hours depending on incremental/full counts, I run 4 
at a time.. Most of the workstations are on 100mbit, it's backing up about 
330 gig total of files.

And yeah, I love BackupPC too :)

Mike


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
Les Mikesell wrote:
 Evren Yurtesen wrote:
 
 The problem is the reads, not writes. It takes long time for BackupPC 
 to figure out if the file should be backed up or not. Backups take 
 very long time even when there are few files which are backed up.
 
 That's only true with rsync and especially rsync fulls where the 
 contents are compared as well as the timestamps. But I thought your 
 complaint was about the hardlinks and data storage.  The transfer speed 
 value doesn't make a lot of sense in the rsync case anyway because you 
 often spend a lot of time computing to avoid having to transfer anything 
 - and you have the entire directory in RAM while doing it which may 
 reduce your disk buffering.  What is wall clock time for a run and is it 
 reasonable for having to read through both the client and server copies?

I am using rsync but the problem is that it still has to go through a 
lot of hard links to figure out if files should be backed up or not.

 But I have to disagree, again :) When you make writes and filesystem 
 is mounted with 'sync' option which is default on FreeBSD then all 
 operations are done sync. Your program must have been doing something 
 between truncate and the actual write operation which caused file to 
 be empty for a period of time. 
 
 Yes, it was writing a small amount to the file and the data write was 
 always deferred with this kind of problem:
 http://www.complang.tuwien.ac.at/anton/sync-metadata-updates.html

You shouldnt be trusting this article. The article is flawed in certain 
ways.

For example it says as demonstrated by soft updates (see below), which 
don't even require fsck upon crashing. but this information is wrong. 
Softupdates means that you can run the fsck in background but you must 
run it on a crash (system does it automatically anyhow).

Also, about the problem in this article, if async updates would have 
been used then the file would have been lost anyhow as the writes would 
have been buffered.

The information given is vague. I am not an expert on this but the 
problem about async updates that the BSD guys are scared of is probably 
not the same kind of inconsistency described here.

In either case you could just enable sync or async updates on UFS to 
change this behaviour. You didnt have to change the filesystem because 
of a deficiency in UFS which is causing the problem.


 The only difference is that FreeBSD uses sync by default because they 
 claim(there is no point to argue if it is true or not, there are 
 enough threads about those if you google :) ) that it is safer while 
 Linux uses async.
 
 It is safer for the file system if you are concerned that fsck can't fix 
 it.  It is not safer for the data of any particular file.  Anyway 
 crashes are pretty much a thing of the past and I protect against most 
 kinds of failures by raid-mirroring to an external drive that is rotated 
 off-site weekly.  I'm running a journaled reiserfs because it is very 
 fast at creating and deleting files, but ext3 with the right options 
 should also work.

If you check namesys.com benchmarks, you will see that they only tested 
reiserfs against ext2/ext3/xfs/jfs and conveniently forgot to test 
against ufs2.

You can see in the end of the page that slight performance increase in 
reiserfs is also bringing twice the cpu usage! (plus extX is faster in 
certain operations even)
http://www.namesys.com/benchmarks.html

Another test results from sun, little bit old (from 2004)
http://www.sun.com/software/whitepapers/solaris10/fs_performance.pdf

I am not saying one is faster over another or anything like that. Sun's 
results do look too good to be true anyhow, they probably tweaked some 
stuff or I dont know, can they lie? wouldnt it be bad reputation?



 Perhaps, but there is a difference if they are moving 10 times or 
 10 times. Where the difference is that the possibility of 
 failure due to mechanical problems increases 1 times.

 No, it doesn't make a lot of difference as long as the drive 
 doesn't overheat.  The head only moves so fast and it doesn't 
 matter if it does it continuously.  However, if your system has 
 sufficient RAM, it will cache and optimize many of the things that 
 might otherwise need an additional seek and access.

 I cant see how you can reach to this conclusion.

 Observation... I run hundreds of servers, many of which are 5 or more 
 years old.  The disk failures have had no correlation to the server 
 activity.

 Hard drive manufacturers disagree with you. They do consider the 
 duty-cycles when calculating the MTBF values.
 http://www.digit-life.com/articles/storagereliability/
 
 Vendor MTBF values are very unrealistic. The trade rags seem to have 
 caught on to this recently: 
 http://www.eweek.com/article2/0%2C1895%2C2105551%2C00.asp
 If you want to know what the vendor really expects, look at the warranty 
 instead.
 

It is a logical fact that more movement = more wear.

Thanks,
Evren


Re: [BackupPC-users] Unable to connect to BackupPC server error

2007-03-27 Thread Ciarlotta, Aaron
That spare hard drive wouldn't happen to be an external USB or Firewire,
would it?


Aaron Ciarlotta
Linux/UNIX Systems Guru

 

-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of
Winston Chan
Sent: Monday, March 26, 2007 8:14 PM
To: backuppc-users@lists.sourceforge.net
Subject: [BackupPC-users] Unable to connect to BackupPC server error

I had been running BackupPC on an Ubuntu computer for several months to
back the computer to a spare hard drive without problem. About the time
I added a new host (Windows XP computer using Samba), I started getting
the following behavior:

BackupPC backs both hosts properly onto the spare hard drive once or
twice after I reboot the Ubuntu server. Then I get a Error: Unable to
connect to BackupPC server error when I attempt to go the web
interface. When I restart BackupPC with /etc/init.d/backuppc restart,
I get a message Can't create LOG file /var/lib/backuppc/log/LOG at
/usr/share/backuppc/bin/BackupPC line 1735.

I have made sure that the backuppc is the owner of
/var/lib/backuppc/log/LOG. It seems the only way to get backuppc to work
again is to reboot the Ubuntu server. I then can see on the web
interface that the last successful backup (of both hosts) occurred 1 or
2 days after the previous reboot. Then backupPC works for 1 or 2 backups
and the cycle starts again.

What is the cause of this and how can I fix it?

Winston




-
Take Surveys. Earn Cash. Influence the Future of IT Join
SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDE
V
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/
The information in this electronic mail message is sender's 
business Confidential and may be legally privileged.  It is 
intended solely for the addressee(s).  Access to this Internet 
electronic mail message by anyone else is unauthorized.  If 
you are not the intended recipient, any disclosure, copying, 
distribution or any action taken or omitted to be taken in 
reliance on it is prohibited and may be unlawful. The sender 
believes that this E-mail and any attachments were free of 
any virus, worm, Trojan horse, and/or malicious code when 
sent. This message and its attachments could have been 
infected during  transmission. By reading the message and 
opening any attachments, the recipient accepts full 
responsibility for taking protective and remedial action about 
viruses and other defects. Travelport Inc. is not liable for any loss 
or damage arising in any way from this message or its 
attachments.


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Evren Yurtesen wrote:

 What is wall clock time for a run and is it 
 reasonable for having to read through both the client and server copies?
 
 I am using rsync but the problem is that it still has to go through a 
 lot of hard links to figure out if files should be backed up or not.

 From the perspective of the backup directories, it doesn't matter 
whether or not there are additional links to a file. It just looks at 
the directory entry to find the file it wants.  It may matter that the 
inodes and file contents end up splattered all over the place because 
they were written at different times, though.

 But I have to disagree, again :) When you make writes and filesystem 
 is mounted with 'sync' option which is default on FreeBSD then all 
 operations are done sync. Your program must have been doing something 
 between truncate and the actual write operation which caused file to 
 be empty for a period of time. 
 Yes, it was writing a small amount to the file and the data write was 
 always deferred with this kind of problem:
 http://www.complang.tuwien.ac.at/anton/sync-metadata-updates.html
 
 You shouldnt be trusting this article. The article is flawed in certain 
 ways.

The fact is that 90+% of the time, my data files were empty after a 
crash while they always appeared to have data when running.  That tells 
me that the truncate happened on disk immediately while the data write 
happened when a buffer was flushed which was consistent with the way the 
bsd sync-metadata was described.

 For example it says as demonstrated by soft updates (see below), which 
 don't even require fsck upon crashing. but this information is wrong. 
 Softupdates means that you can run the fsck in background but you must 
 run it on a crash (system does it automatically anyhow).

This was before softupdates was an option.  I don't think reiserfs was 
around then either.

 Also, about the problem in this article, if async updates would have 
 been used then the file would have been lost anyhow as the writes would 
 have been buffered.

But then you'd have about a 99% chance of still having the previous 
contents which would have been much more useful than an empty file.

 In either case you could just enable sync or async updates on UFS to 
 change this behaviour. You didnt have to change the filesystem because 
 of a deficiency in UFS which is causing the problem.

I just got tired of the freebsd people telling me that their way was 
better and it was correct behavior to lose my data and moved on.

 The only difference is that FreeBSD uses sync by default because they 
 claim(there is no point to argue if it is true or not, there are 
 enough threads about those if you google :) ) that it is safer while 
 Linux uses async.
 It is safer for the file system if you are concerned that fsck can't fix 
 it.  It is not safer for the data of any particular file.  Anyway 
 crashes are pretty much a thing of the past and I protect against most 
 kinds of failures by raid-mirroring to an external drive that is rotated 
 off-site weekly.  I'm running a journaled reiserfs because it is very 
 fast at creating and deleting files, but ext3 with the right options 
 should also work.
 
 If you check namesys.com benchmarks, you will see that they only tested 
 reiserfs against ext2/ext3/xfs/jfs and conveniently forgot to test 
 against ufs2.
 
 You can see in the end of the page that slight performance increase in 
 reiserfs is also bringing twice the cpu usage! (plus extX is faster in 
 certain operations even)
 http://www.namesys.com/benchmarks.html

There's a benchmark called 'postmark' that is the best I've seen for 
showing how well creating/deleting a lot of files works.  A few years 
ago when I changed last, reiserfs was the best.  It seems to have 
disappeared from it's original location from NetApp, but the last time I 
mentioned it someone said it was packaged for Debian so the source is 
still available.  I'll dig it up again the next time I am interested in 
changing (maybe when zfs looks mature).

 Perhaps, but there is a difference if they are moving 10 times or 
 10 times. 
 
 It is a logical fact that more movement = more wear.

Yes, but more wear doesn't matter until you exceed the design limits. 
Worry about your air conditioner instead - it has more effect on the 
drive life. I've had as much trouble with drives that have been powered 
off for much of their lives as ones that stay busy.

-- 
   Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net

Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Carl Wilhelm Soderstrom
On 03/27 11:48 , Les Mikesell wrote:
 Worry about your air conditioner instead - it has more effect on the 
 drive life. I've had as much trouble with drives that have been powered 
 off for much of their lives as ones that stay busy.

as an interesting data point, I have a 512MB drive that ran my firewall for
several years, and before that was in a desktop machine at a government
office. The date on it is 1993; and it still works just fine, despite what
I'm guessing is fairly regular and continuous use since that time.

I would really like to see hard drives made to be more reliable, rather than
just bigger. (Where the mfg. actually touts the reliability features more
than the size). Maxtor sort of tried that a while ago with their
single-platter 15GB drives; but AFAIK there wasn't anything special about
those other than fewer platters and heads (no better bearings,
gas-absorption material, etc). We've long since passed the point where big
drives are necessary (at least for OS partitions on Linux boxen). I want
reliable storage.

to that end, this is kind of an attractive product:
http://www.elandigitalsystems.com/adapter/p312.php

we've had good luck with these units:
http://www.mach5products.com/PRODCF/CFproducts.htm

My ~1TB BackupPC server is using less than 1GB on its OS drive (the backuppc
data is all on a separate drive array, for ease of management). While
consumer-grade CF cards are pretty crappy things compared to the
military-grade solid-state IDE or SCSI disks that places like M-Systems
sells; they are cheap enough and big enough to compete with spinny storage
when the space requirements take second place to the reliability of no
moving parts. (I've seen CF cards die... I'm not sure that a spinny drive
wouldn't have died in the same conditions tho).

-- 
Carl Soderstrom
Systems Administrator
Real-Time Enterprises
www.real-time.com

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread David Rees
On 3/27/07, Les Mikesell [EMAIL PROTECTED] wrote:
 Evren Yurtesen wrote:

  What is wall clock time for a run and is it
  reasonable for having to read through both the client and server copies?
 
  I am using rsync but the problem is that it still has to go through a
  lot of hard links to figure out if files should be backed up or not.

Evren, I didn't see that you mentioned a wall clock time for your
backups? I want to know how many files are in a single backup, how
much data is in that backup and how long it takes to perform that
backup.

  From the perspective of the backup directories, it doesn't matter
 whether or not there are additional links to a file. It just looks at
 the directory entry to find the file it wants.  It may matter that the
 inodes and file contents end up splattered all over the place because
 they were written at different times, though.

Yep, Lee is right here. Unless BSD handles hard-links in some crazy manner.

  If you check namesys.com benchmarks, you will see that they only tested
  reiserfs against ext2/ext3/xfs/jfs and conveniently forgot to test
  against ufs2.
 
  You can see in the end of the page that slight performance increase in
  reiserfs is also bringing twice the cpu usage! (plus extX is faster in
  certain operations even)
  http://www.namesys.com/benchmarks.html

When your overall speed is limited by the speed of your disks and your
CPU spends all it's time twiddling it's thumbs waiting for disk, who
cares if CPU doubles and still spends 90% of it's time waiting as long
as it gets the job done faster?

To summarize Evran's setup:

FreeBSD, 250MB ram, CPU unknown, 1 7200rpm Seagate ATA disk
Filesystem: ufs2, sync, noatime
Pool is 17.08GB comprising 760528 files (avg file size ~22KB)
BackupPC reports backup speeds between 0.06 - 0.22 MB/s
Total backup time per host: Unknown
CPU is 99% idle during backups
Disk shows ~75 IO/s during load and low transfer rate
Says even small backup w/small number of files is slow.

Can you try mounting the backup partition async so we can see if it
really is read performance or write performance that is killing backup
performance?

I would also highly recommend that you limit backups to 1 concurrent
backup at a time.

I must wonder if ufs2 is really bad at storing inodes on disk...

BTW, how does BackupPC calculate speed? I think it calculates backup
speed by reporting files transferred over time, so if you don't have
many files that change, won't BackupPC report a very low backup speed.

-Dave

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Craig Barratt
Evren writes:

 I use rsync with 3.0.0 but it was the same speed with 2.x.x versions.
 
 $Conf{IncrLevels} = [1];
 
 I wonder, since I have:
 
 $Conf{IncrKeepCnt} = 6;
 
 Wouldnt it make more sense to use this?:
 
 $Conf{IncrLevels}  = [1, 2, 3, 4, 5, 6];
 
 or does this make things even more slower?

Yes, that would make it slower since several incrementals have
to be merged (ie: more server disk seeks for every file) to get
the latest baseline, which in 3.x is used as the starting point
for each rsync backup.

Craig

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread David Rees
On 3/27/07, David Rees [EMAIL PROTECTED] wrote:
 Can you try mounting the backup partition async so we can see if it
 really is read performance or write performance that is killing backup
 performance?

 I must wonder if ufs2 is really bad at storing inodes on disk...

I went and did some research on ufs filesystem performance and found this paper:
http://www.bsdcan.org/2006/papers/FilesystemPerformance.pdf

There appears to be 4 different mount options related to data integrity:

sync: slowest, all writes synchronous
noasync: (default?) data asynchronous, metadata synchronous
soft updates: dependency ordering of writes to ensure on-disk consistency
async: fastest, all writes asynchronous

noasync seems to be the default. Evran, can you confirm that your
filesystem is mounted this way?

On Linux using ext3, the default mount option is data=ordered which
should be similar to soft updates in terms of performance. If you can
mount your backup partition in soft updates mode that should perform
best, much better than the default noasync mode for the type of writes
BackupPC does.

I wouldn't recommend mounting a partition async permanently because of
the risk to file system integrity.

-Dave

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Carl Wilhelm Soderstrom wrote:

 I would really like to see hard drives made to be more reliable, rather than
 just bigger.

I'm not sure that can be improved enough to matter.  A failure of an 
inexpensive part once in five years isn't a big deal other than the side 
effects it might cause, and unless they can be 100% reliable (probably 
impossible) you have to be prepared for those side effects anyway.

  I want reliable storage.

Run mirrored hot-swap drives. If one dies, replace it at some convenient 
time, sync it up and keep going.  I have one machine last booted in 
mid-2003 that has had a couple of it's drives replaced that way.

-- 
   Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Carl Wilhelm Soderstrom
On 03/27 01:40 , Les Mikesell wrote:
 Carl Wilhelm Soderstrom wrote:
  I would really like to see hard drives made to be more reliable, rather than
  just bigger.
 I'm not sure that can be improved enough to matter.  A failure of an 
 inexpensive part once in five years isn't a big deal other than the side 
 effects it might cause, and unless they can be 100% reliable (probably 
 impossible) you have to be prepared for those side effects anyway.

on a small scale, perhaps. 
when you get to an organization the size of (TBS|Hotmail|Yahoo)tho; it may
make sense to spend a little more money in order to spend less labor
swapping parts.

the market seems to agree with you tho... which is a pretty good indictment
of central planning notions (i.e. Me) vs. what really is efficient.

   I want reliable storage.
 Run mirrored hot-swap drives. If one dies, replace it at some convenient 
 time, sync it up and keep going.  I have one machine last booted in 
 mid-2003 that has had a couple of it's drives replaced that way.

Your point is well-taken. I do keep in mind tho that I've seen multiple
drives fail simultaneously or in rapid succession, and that the process of
replacing drives costs time and effort. It is not as trivial as you might
think, once you factor in the time to detect the failure, find the
replacement (if you can), replace the drive (which may involve removing the
old drive from the carrier and replacing it with the new one), and make sure
it's working properly. In an ideal world it's 5 minutes of work, in the real
world I usually expect to lose an hour or two dealing with a failed drive.
This can be accounted as several hundred dollars of lost productivity in
many cases; so it's worthwhile to spend more money on a better-quality drive
sometimes.

-- 
Carl Soderstrom
Systems Administrator
Real-Time Enterprises
www.real-time.com

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Timing out full backups

2007-03-27 Thread Michael Mansour
Hi Les,

Apologies for the late reply.

 Michael Mansour wrote:
  Hi,
  
  Hi,
 
  I have the following host summary:
 
  server1 username1   20.63.360.772   10.4   
   
  10.4idledone server2   username1   20.63.23
  0.743   10.4 
 10.4idledone server3username2   20.68.52
  1.894   18.6
  18.6   idledone server4username2   20.63.01
  1.415   10.4 
 10.4idledone server5username2   20.63.38
  1.244   18.6
  18.6   idledone server6username2   20.62.86
  1.445   10.4 
 10.4idledone
 
  This is on a test BackupPC server still using Beta 3.0 (I'll soon 
  build a production BackupPC server to replace this box).
 
  I'm wondering why the full backups numbered 2 are not going back 
  down to 1 to free up some space on the server?
 
  In the glbal Schedule, I have the following:
 
  FullPeriod: 6.97
  FullKeepCnt: 1
  FullKeepCntMin: 1
  FullAgeMax: 7
 
  and it's my understanding that backuppc should cycle those full 
  backups above to trash because of the above globals yes?
  
  Did anyone have any ideas on this?
 
 Full backups that incrementals are based from are not deleted until 
 after the corresponding incremental runs have been removed, and your 
 last full will not be removed until the next one is complete.

I understand, but the stats are:

There are 6 hosts that have been backed up, for a total of: 

10 full backups of total size 44.55GB (prior to pooling and compression), 
15 incr backups of total size 8.44GB (prior to pooling and compression). 

I have servers which have multiple full backups and incrementals, but neither
is removed.

Is there a way I can just manually delete some of the backups (incrementals
and fulls) to start this process again? as I can't seem to do anything while
the disks are full and the timeout processes are not working.

Thanks.

Michael.


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Timing out full backups

2007-03-27 Thread Les Mikesell
Michael Mansour wrote:

 I have servers which have multiple full backups and incrementals, but neither
 is removed.

I don't think they are removed until the replacements are completed.

 Is there a way I can just manually delete some of the backups (incrementals
 and fulls) to start this process again? as I can't seem to do anything while
 the disks are full and the timeout processes are not working.

Yes, you can go to the pc directory and 'rm -rf backup_number' for what 
you want to delete.  The space won't actually be released until 
BackupPC_nightly runs and then only for files that are not linked by 
other backup runs.  Some of your web displays will be wrong for a while 
after deleting things this way but it will eventually correct itself.

-- 
  Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Timing out full backups

2007-03-27 Thread Michael Mansour
Hi Les,

 Michael Mansour wrote:
 
  I have servers which have multiple full backups and incrementals, but 
  neither
  is removed.
 
 I don't think they are removed until the replacements are completed.

  Is there a way I can just manually delete some of the backups (incrementals
  and fulls) to start this process again? as I can't seem to do anything while
  the disks are full and the timeout processes are not working.
 
 Yes, you can go to the pc directory and 'rm -rf backup_number' for 
 what you want to delete.  The space won't actually be released until 
 BackupPC_nightly runs and then only for files that are not linked by 
 other backup runs.  Some of your web displays will be wrong for a 
 while after deleting things this way but it will eventually correct itself.

Ok, I have started deleting and will kick off some manual incrementals 
afterwards.

Thanks.

Michale.

 -- 
   Les Mikesell
 [EMAIL PROTECTED]
--- End of Original Message ---


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Timing out full backups

2007-03-27 Thread Craig Barratt
Michael writes:

  Michael Mansour wrote:
  
   I have servers which have multiple full backups and incrementals, but 
   neither
   is removed.
  
  I don't think they are removed until the replacements are completed.
 
   Is there a way I can just manually delete some of the backups 
   (incrementals
   and fulls) to start this process again? as I can't seem to do anything 
   while
   the disks are full and the timeout processes are not working.
  
  Yes, you can go to the pc directory and 'rm -rf backup_number' for 
  what you want to delete.  The space won't actually be released until 
  BackupPC_nightly runs and then only for files that are not linked by 
  other backup runs.  Some of your web displays will be wrong for a 
  while after deleting things this way but it will eventually correct itself.
 
 Ok, I have started deleting and will kick off some manual incrementals 
 afterwards.

You should be *really* sure you know what you are doing when
you manually delete backups.  At a miminum you should remove the
corresponding line in the backups file.

Your stats suggest you have 1 or 2 fulls per host.  That's typically
the minimum number, since subsequent incrementals require the baseline
full to be kept.  For example, even if you specify $Conf{FullKeepCnt} = 1,
you will often have two fulls because the subsequent incrementals need
it kept:

 oldest -   #0 Full
 #1 Incr
 #2 Incr
 #4 Full
 #5 Incr
 #6 Incr
 #7 Incr
 newest -   #8 Incr

 
This is a very common situation.  If you remove full #0 then you have
rendered backups #1 and #2 invalid and useless.

The only three options to reduce the number of backups in this case is:

- remove incrementals #8, #7, #6, #5 (in that order)

- remove incrementals #2, #1 (in that order)

- remove full #0, and incr #1 and #2 (all or nothing).

I'd really recommend against manually remove backups unless you either
really know what you are doing or you are starting over after running
a temporary or test setup.

You can accomplish the same effect as above by setting $Conf{IncrKeepCnt}
to 4 or less.

But even then there will still be cases where there are 2 full backups
kept.  The additional storage is much less than you expect because of
hardlinking.

Craig

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Carl Wilhelm Soderstrom wrote:

 I would really like to see hard drives made to be more reliable, rather than
 just bigger.
 I'm not sure that can be improved enough to matter.  A failure of an 
 inexpensive part once in five years isn't a big deal other than the side 
 effects it might cause, and unless they can be 100% reliable (probably 
 impossible) you have to be prepared for those side effects anyway.
 
 on a small scale, perhaps. 
 when you get to an organization the size of (TBS|Hotmail|Yahoo)tho; it may
 make sense to spend a little more money in order to spend less labor
 swapping parts.

I think before you get to that that scale you'd be network booting 
diskless and mostly disposable motherboards.  You still have to deal 
with the drives somewhere, but using larger ones instead of large 
numbers of them.

 the market seems to agree with you tho... which is a pretty good indictment
 of central planning notions (i.e. Me) vs. what really is efficient.

I find complete-system redundancy with commodity boxes to be cheaper and 
more efficient than using military strength parts that cost 10x as much 
and fail half as often.  In our operations the most common failure is in 
software anyway since we try to push out new capabilities as fast as 
possible so I like lots of redundancy and the ability to run multiple 
versions of things concurrently.  I could see where that would be 
different for operations where everything had to be in a single 
database, though.

   I want reliable storage.
 Run mirrored hot-swap drives. If one dies, replace it at some convenient 
 time, sync it up and keep going.  I have one machine last booted in 
 mid-2003 that has had a couple of it's drives replaced that way.
 
 Your point is well-taken. I do keep in mind tho that I've seen multiple
 drives fail simultaneously or in rapid succession, and that the process of
 replacing drives costs time and effort.

Yes, your building might catch fire or be flooded too. Those are 
probably more likely than multiple cheap drives failing simultaneously 
from some cause that wouldn't also kill better drives. You need offsite 
backups to cover these unlikely events anyway.

 It is not as trivial as you might
 think, once you factor in the time to detect the failure, find the
 replacement (if you can), replace the drive (which may involve removing the
 old drive from the carrier and replacing it with the new one), and make sure
 it's working properly. In an ideal world it's 5 minutes of work, in the real
 world I usually expect to lose an hour or two dealing with a failed drive.

The real killer is the time planning out the replacement operation so 
the only time wasted is the one person who does the work.  In the 
whole-system scenario where the systems are load balanced anyway, 
yanking a machine out isn't anything special and you don't even need the 
hot-swap drive case.

  This can be accounted as several hundred dollars of lost
  productivity in many cases; so it's worthwhile to spend more money
  on a better-quality drive sometimes.

In practice I don't see any relationship between price and reliability. 
  I'm inclined to think that they all come off the same production line 
these days.

-- 
   Les Mikesell
[EMAIL PROTECTED]

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


[BackupPC-users] drive quality (was: Re: very slow backup speed)

2007-03-27 Thread Carl Wilhelm Soderstrom
On 03/27 02:49 , Les Mikesell wrote:
 I find complete-system redundancy with commodity boxes to be cheaper and 
 more efficient than using military strength parts that cost 10x as much 
 and fail half as often.  

That's Google's solution, and it works for them. Nothing fancy... rebooting
servers means a dude with a broomstick walking around pushing power buttons.

 The real killer is the time planning out the replacement operation so 
 the only time wasted is the one person who does the work.  In the 
 whole-system scenario where the systems are load balanced anyway, 
 yanking a machine out isn't anything special and you don't even need the 
 hot-swap drive case.

it's a nice ideal circumstance. :)
In practice I've found that clustering systems often increases
administrative headaches exponentially. There's a lot more that *can* go
wrong and therefore *does* go wrong with a cluster. There's a place for
them, like everything else... but they are far from a panacea.

 In practice I don't see any relationship between price and reliability. 
   I'm inclined to think that they all come off the same production line 
 these days.

some are still better than others... how much better, I don't know. I've
seen some really expensive hardware go belly-up at times too.

Here's what I did with some of the aforementioned bad  expensive hardware:
http://www.redchrome.org/shooting_computers/shooting_computers_index.html
(the Sparc 10, specifically)

-- 
Carl Soderstrom
Systems Administrator
Real-Time Enterprises
www.real-time.com

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
David Rees wrote:
 On 3/27/07, Les Mikesell [EMAIL PROTECTED] wrote:
 Evren Yurtesen wrote:

 What is wall clock time for a run and is it
 reasonable for having to read through both the client and server copies?
 I am using rsync but the problem is that it still has to go through a
 lot of hard links to figure out if files should be backed up or not.
 
 Evren, I didn't see that you mentioned a wall clock time for your
 backups? I want to know how many files are in a single backup, how
 much data is in that backup and how long it takes to perform that
 backup.

I sent the status of the backups earlier today to mailing list?

# Pool is 17.08GB comprising 760528 files and 4369 directories (as of 
3/27 05:54),
# Pool hashing gives 38 repeated files with longest chain 6,
# Nightly cleanup removed 10725 files of size 0.40GB (around 3/27 05:54),
# Pool file system was recently at 10% (3/27 22:40), today's max is 10% 
(3/27 01:00) and yesterday's max was 10%.


 *  16 full backups of total size 72.16GB (prior to pooling and 
compression),
 * 24 incr backups of total size 13.60GB (prior to pooling and 
compression).


This is from 1 machine.

 Totals  Existing Files  New Files
Backup# Type#Files  Size/MB MB/sec  #Files  Size/MB 
#Files  Size/MB
112 full149290  1957.8  0.10149251  1937.9  460120.7
224 full151701  2022.8  0.09151655  2004.7  100 18.1
238 full152170  2099.5  0.06152131  2081.6  115 17.9
244 incr214 48.90.00165 22.978  26.0
245 full152228  2095.2  0.06152177  2076.9  108 18.3
246 incr118 17.30.0076  0.2 69  17.1
247 incr159 21.40.00111 3.1 75  18.4
248 incr181 22.10.00132 2.5 79  19.7
249 incr186 24.00.00146 7.6 54  16.4
250 incr206 25.50.00159 6.7 70  18.8



  From the perspective of the backup directories, it doesn't matter
 whether or not there are additional links to a file. It just looks at
 the directory entry to find the file it wants.  It may matter that the
 inodes and file contents end up splattered all over the place because
 they were written at different times, though.
 
 Yep, Lee is right here. Unless BSD handles hard-links in some crazy manner.

I dont know if the problem is hard links. This is not a FreeBSD or Linux 
problem. It exists on both. Just that people using ultra fast 5 disk 
raid 5 setups are seeing 2mbytes/sec transfer rate means that backuppc 
is very very inefficient.

For example this guy is using Linux (problem is OS independent)
http://forum.psoft.net/showpost.php?p=107808postcount=16


 If you check namesys.com benchmarks, you will see that they only tested
 reiserfs against ext2/ext3/xfs/jfs and conveniently forgot to test
 against ufs2.

 You can see in the end of the page that slight performance increase in
 reiserfs is also bringing twice the cpu usage! (plus extX is faster in
 certain operations even)
 http://www.namesys.com/benchmarks.html
 
 When your overall speed is limited by the speed of your disks and your
 CPU spends all it's time twiddling it's thumbs waiting for disk, who
 cares if CPU doubles and still spends 90% of it's time waiting as long
 as it gets the job done faster?

Whatever, this is not the problem here.The fact is that, according to 
reiserfs developers reiserfs is more or less the same speed with ext2. I 
dont think the problem is related to any filesystem as it occurs on both 
Linux and FreeBSD

 To summarize Evran's setup:
 
 FreeBSD, 250MB ram, CPU unknown, 1 7200rpm Seagate ATA disk
 Filesystem: ufs2, sync, noatime
 Pool is 17.08GB comprising 760528 files (avg file size ~22KB)
 BackupPC reports backup speeds between 0.06 - 0.22 MB/s
 Total backup time per host: Unknown
 CPU is 99% idle during backups
 Disk shows ~75 IO/s during load and low transfer rate
 Says even small backup w/small number of files is slow.
 
 Can you try mounting the backup partition async so we can see if it
 really is read performance or write performance that is killing backup
 performance?
 
 I would also highly recommend that you limit backups to 1 concurrent
 backup at a time.
 
 I must wonder if ufs2 is really bad at storing inodes on disk...

On Linux with raid setup with async io etc. people are getting slightly 
better results. I think ufs2 is just fine. I wonder if there is 
something in my explanations...The problem is backuppc. People are 
getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives, 
using Linux. It is a miracle that backup even finishes in 24 hours using 
a standart ide drive.


 BTW, how does BackupPC calculate speed? I think it calculates backup
 speed by reporting files transferred over time, so if you don't have
 many files that change, won't BackupPC report a very low backup speed.

This is like the 'Contact' 

Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Jason Hughes

Evren Yurtesen wrote:

 Totals  Existing Files  New Files
Backup# Type#Files  Size/MB MB/sec  #Files  Size/MB 
#Files  Size/MB
245 full152228  2095.2  0.06152177  2076.9  108 18.3
246 incr118 17.30.0076  0.2 69  17.1
  


Can you post the duration that these backups took?  All that these stats 
tell us is how much of your data is churning, and how big your average 
file size is.


I dont know if the problem is hard links. This is not a FreeBSD or Linux 
problem. It exists on both. Just that people using ultra fast 5 disk 
raid 5 setups are seeing 2mbytes/sec transfer rate means that backuppc 
is very very inefficient.


For example this guy is using Linux (problem is OS independent)
http://forum.psoft.net/showpost.php?p=107808postcount=16

  


Er, RAID-5 is slower than a single disk except on sustained reads, which 
is not the typical case with BackupPC.  I only use a single disk and get 
between 0.5mb/s and 10mb/s, depending on the client and its file 
size/file count distribution.  The question is whether other people have 
gotten the system to work correctly under the same or worse conditions, 
and already that has been answered in the affirmative.  Be prepared to 
accept that the software isn't broken just because you haven't gotten it 
working to your satisfaction in your specific situation...  It could be 
rsync, UFS, BSD, or any number of other factors that are causing you grief.


Whatever, this is not the problem here.The fact is that, according to 
reiserfs developers reiserfs is more or less the same speed with ext2. I 
dont think the problem is related to any filesystem as it occurs on both 
Linux and FreeBSD
  


Your argument lacks logic.  If a filesystem can be configured to be slow 
on multiple OS's, does that mean BackupPC is failing to do its job?  No, 
it means multiple people have managed to set up a system that performs 
badly using it.  That's not so uncommon.  BackupPC does not exist in a 
vacuum: its performance is sensitive to the environment of the server 
and its clients.  Many people are using it without issue, right here on 
this very list.  The question you should be asking is, what makes your 
system perform badly?  Start by picking out things that you do 
differently, or that definitely affect performance.  Transport protocol 
(rsync, tar).  UFS.  BSD.   Sync on your filesystem.  Start changing 
those one at a time and measure the performance.


On Linux with raid setup with async io etc. people are getting slightly 
better results. I think ufs2 is just fine. I wonder if there is 
something in my explanations...The problem is backuppc. People are 
getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives, 
using Linux. It is a miracle that backup even finishes in 24 hours using 
a standart ide drive.


  


If you want people to help you, it's probably best if you refrain from 
blaming the program until you have proof.  So far, you have argued with 
people who provide evidence that the system works fine.  It puts people 
on the defensive, and you may find people less willing to help you in 
the future.  We do know BackupPC behaves well on other file systems and 
operating systems... maybe UFS or BSD is doing something poorly--how 
would you know they aren't?  We have less data points to draw from there.


I suspect it has a lot more to do with what the MB/s stats really mean.  
Maybe Craig can give a precise definition?


This is like the 'Contact' movie. The sphere took 30 seconds to download 
but there were 18 hours of recording. If what you said was true and 
backuppc would be backing up very small amount of files and skipping 
most, then backups would probably take less time than 2-4 hours each.


  
With rsync in incremental backups, it still has to check the metadata 
for each file to determine the changed file set.  With millions of 
files, it will take a while.  If your host or server is low on memory 
for any reason, this may bog it down and start vm swapping.  I would 
recommend trying out tar to see if the protocol behavior matters.  Try 
different mount options that are for higher performance.  Others have 
made similar suggestions as well.


Good luck,
JH

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Brien Dieterle




Evren Yurtesen wrote:

  David Rees wrote:
  
  
On 3/27/07, Les Mikesell [EMAIL PROTECTED] wrote:


  Evren Yurtesen wrote:

  
  

  What is wall clock time for a run and is it
reasonable for having to read through both the client and server copies?
  

I am using rsync but the problem is that it still has to go through a
lot of hard links to figure out if files should be backed up or not.

  

Evren, I didn't see that you mentioned a wall clock time for your
backups? I want to know how many files are in a single backup, how
much data is in that backup and how long it takes to perform that
backup.

  
  
I sent the status of the backups earlier today to mailing list?

# Pool is 17.08GB comprising 760528 files and 4369 directories (as of 
3/27 05:54),
# Pool hashing gives 38 repeated files with longest chain 6,
# Nightly cleanup removed 10725 files of size 0.40GB (around 3/27 05:54),
# Pool file system was recently at 10% (3/27 22:40), today's max is 10% 
(3/27 01:00) and yesterday's max was 10%.


 *  16 full backups of total size 72.16GB (prior to pooling and 
compression),
 * 24 incr backups of total size 13.60GB (prior to pooling and 
compression).


This is from 1 machine.

  	 Totals  	 Existing Files  	 New Files
Backup# 	Type 	#Files 	Size/MB 	MB/sec 	#Files 	Size/MB 	#Files 	Size/MB
112 	full 	149290 	1957.8 	0.10 	149251 	1937.9 	4601 	20.7
224 	full 	151701 	2022.8 	0.09 	151655 	2004.7 	100 	18.1
238 	full 	152170 	2099.5 	0.06 	152131 	2081.6 	115 	17.9
244 	incr 	214 	48.9 	0.00 	165 	22.9 	78 	26.0
245 	full 	152228 	2095.2 	0.06 	152177 	2076.9 	108 	18.3
246 	incr 	118 	17.3 	0.00 	76 	0.2 	69 	17.1
247 	incr 	159 	21.4 	0.00 	111 	3.1 	75 	18.4
248 	incr 	181 	22.1 	0.00 	132 	2.5 	79 	19.7
249 	incr 	186 	24.0 	0.00 	146 	7.6 	54 	16.4
250 	incr 	206 	25.5 	0.00 	159 	6.7 	70 	18.8



  
  

   From the perspective of the backup directories, it doesn't matter
whether or not there are additional links to a file. It just looks at
the directory entry to find the file it wants.  It may matter that the
inodes and file contents end up splattered all over the place because
they were written at different times, though.
  

Yep, Lee is right here. Unless BSD handles hard-links in some crazy manner.

  
  
I dont know if the problem is hard links. This is not a FreeBSD or Linux 
problem. It exists on both. Just that people using ultra fast 5 disk 
raid 5 setups are seeing 2mbytes/sec transfer rate means that backuppc 
is very very inefficient.
  

Well, I'd like to point out that I am using tar over nfs over a bonded
100mbit(x2) link. NFS is terrible for small file access. However, we
have an xserve that I use xtar with, over ssh-- I get 5MB/s for that
(which is the best I've seen). My other rsync clients get up to 4.2
MB/sec. So, I think the 25% [EMAIL PROTECTED]/s indicates I should never
expect more than 10MB/s.

Considering everything backuppc has to do with each file as it
receives them, without writing them to disk first, I don't think we
can expect it to be very fast. However, it is absurdly efficient with
space. You can't have everything :-). Complexity, speed, efficiency;
pick any two? 

brien





-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Evren Yurtesen wrote:

 #  Type  #Files  Size/MB  MB/sec  #Files  Size/MB  #Files Size/MB
 245 full  152228  2095.2   0.06   152177  2076.910818.3

 On Linux with raid setup with async io etc. people are getting slightly 
 better results. I think ufs2 is just fine. I wonder if there is 
 something in my explanations...The problem is backuppc. People are 
 getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives, 
 using Linux. It is a miracle that backup even finishes in 24 hours using 
 a standart ide drive.

2MB/sec isn't bad when handling a lot of files (try unpacking a tar with 
  hundreds of thousands of little files to see).  The problem is that 
you are getting .06 on a drive that is capable of running a couple of 
sessions at 1.5MB/sec or so. I'd try to get up to normal speed before 
claiming that the software needs to be fixed to get more.  More RAM will 
probably make a huge difference, but first are you sure that your IDE 
controller and cable are a match for the drive and that your OS is using 
an efficient DMA mode?  I've seen even some of the 80-pin cables have a 
problem that would make things shift down to 33Mhz pio mode which will 
kill you.  Does freebsd log the IDE mode detected and have some way to 
test throughput?

 BTW, how does BackupPC calculate speed? I think it calculates backup
 speed by reporting files transferred over time, so if you don't have
 many files that change, won't BackupPC report a very low backup speed.
 
 This is like the 'Contact' movie. The sphere took 30 seconds to download 
 but there were 18 hours of recording. If what you said was true and 
 backuppc would be backing up very small amount of files and skipping 
 most, then backups would probably take less time than 2-4 hours each.

If you look at the 'duration' in the backup summary and the Size/MB in 
the lower  File Size summary you can compute your own rate based on what 
you are backing up.  For fulls at least this seems to be based on the 
real target size, not what rsync had to transfer.

-- 
  Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Les Mikesell
Evren Yurtesen wrote:

 Raid5 doesn't distribute disk activity - it puts the drives in 
 lockstep and is slower than a single drive, especially on small writes 
 where it has to do extra reads to re-compute parity on the existing data.
 
 I am confused, when a write is done the data is distributed in the disks 
 depending on the stripe size you are using. When you start reading the 
 file, you are reading from 5 different disks. So you get way better 
 performance for sure on reads.

The stripe effect only comes into play on files large enough to span 
them and not at all for directory/inode accesses which is most of what 
you are doing. Meanwhile you have another head tied up checking the 
parity and for writes of less than a block you have to read the existing 
contents before the write to re-compute the parity.

-- 
   Les Mikesell
[EMAIL PROTECTED]


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread David Rees
On 3/27/07, Evren Yurtesen [EMAIL PROTECTED] wrote:
 David Rees wrote:
  Evren, I didn't see that you mentioned a wall clock time for your
  backups? I want to know how many files are in a single backup, how
  much data is in that backup and how long it takes to perform that
  backup.

 I sent the status of the backups earlier today to mailing list?

Still missing wall-clock time. Though we can extrapolate that it may
be taking over 6 hours for a full backup of the system below. Is that
correct?

 This is from 1 machine.

  Totals  Existing Files  New Files
 Backup# Type#Files  Size/MB MB/sec  #Files  Size/MB   
   #Files  Size/MB
 112 full149290  1957.8  0.10149251  1937.9  460120.7

 I dont know if the problem is hard links. This is not a FreeBSD or Linux
 problem. It exists on both. Just that people using ultra fast 5 disk
 raid 5 setups are seeing 2mbytes/sec transfer rate means that backuppc
 is very very inefficient.

As mentioned earlier, RAID 5 is horrible for random small read/write
performance. It is often worse than a single disk.

But still, I have a client which has 1.5 million files and 80-100GB of
data. A full backup takes 4-6 hours which is reasonable. Full backups
average 4.5-5MB/s.

 For example this guy is using Linux (problem is OS independent)
 http://forum.psoft.net/showpost.php?p=107808postcount=16

Your transfer rates are at least 5x slower than his and 10x slower
than what most people here are getting. There is something wrong
specifically with your setup. I suspect that if your backups were
going 5x faster you'd be happy and 10x faster you'd be very happy.

 On Linux with raid setup with async io etc. people are getting slightly
 better results. I think ufs2 is just fine. I wonder if there is
 something in my explanations...The problem is backuppc. People are
 getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives,
 using Linux. It is a miracle that backup even finishes in 24 hours using
 a standart ide drive.

With ext2 the default is async IO. With ext3 (the default system now)
the default is ordered which is similar to BSD's soft updates.

 This is like the 'Contact' movie. The sphere took 30 seconds to download
 but there were 18 hours of recording. If what you said was true and
 backuppc would be backing up very small amount of files and skipping
 most, then backups would probably take less time than 2-4 hours each.

Using the rsync transfer method, it does require at least a stat of
every single file to be backed up. For 150k files and 2GB of data,
you'd really expect this to be done within a hour with nearly any
machine.

Are you using the --checksum-seed option? That will significantly
speed up (the 3rd and later) backups as well.

Are you also sure the backup server isn't swapping as well? Can you
get us some good logs from vmstat or similar during a backup?

I also suspect that there is something else in this case slowing your
machine down. Unless you give us the information to help track it
down, we will not be able to figure it out. I feel like I am pulling
teeth.

-Dave

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] RSync v. Tar

2007-03-27 Thread Holger Parplies
Hi,

Les Mikesell wrote on 27.03.2007 at 01:03:32 [Re: [BackupPC-users] RSync v. 
Tar]:
 Jesse Proudman wrote:
  I've got one customer who's server has taken 3600 minutes to  
  backup.   77 Gigs of Data.  1,972,859 small files.  Would tar be  
  better or make this faster?  It's directly connected via 100 Mbit to  
 ^^^
  the backup box.
^^
 If the files don't change frequently, tar incremental runs will be much 
 faster because they are based only on the target timestamps while rsync 
will load the entire directory at both ends and compare them.

if you ask me, regardless of how your data changes, tar is the way to go,
not rsync, especially for *full* backups. With a direct 100 MBit connection,
there's not much point in spending (lots of) CPU time for saving bandwidth
- not with 2 million files. rsync is good for low bandwidth connections,
where the link severely limits the transfer and speeding it up makes a real
difference. In your case, your link speed is in the same order of magnitude
as your disk I/O performance (considering our favorite topic, the seek times
on the pool file system, the network link may in fact not even be the limiting
factor - it clearly isn't, as 77 GB would take slightly more than 2 hours to
transfer over a 100 MBit link, and rsync is not making it go faster than
that ;-).

rsync has additional benefits concerning finding and backing up new (or
moved) files with old timestamps and deleted files on *incremental* backups,
but keeping the list of 2 million files in memory will probably be a problem,
as it possibly was with your full (?) backup (how much memory do the BackupPC
server and the backed up host have?). 

(Les: if the files *do* change frequently, there's even less speedup to get
 from using rsync. Only frequent metadata changes without file content
 changes would give rsync an advantage - assuming it is faster to figure out
 that the file is identical than to simply send it over the network.)

Regards,
Holger

P.S.: rsync checksum caching *might* make a difference starting from the
  third backup, but I read that it's less improvement than one might
  expect.

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread John Pettitt
Les Mikesell wrote:
 Evren Yurtesen wrote:
   
 Raid5 doesn't distribute disk activity - it puts the drives in 
 lockstep and is slower than a single drive, especially on small writes 
 where it has to do extra reads to re-compute parity on the existing data.
   
 I am confused, when a write is done the data is distributed in the disks 
 depending on the stripe size you are using. When you start reading the 
 file, you are reading from 5 different disks. So you get way better 
 performance for sure on reads.
 

 The stripe effect only comes into play on files large enough to span 
 them and not at all for directory/inode accesses which is most of what 
 you are doing. Meanwhile you have another head tied up checking the 
 parity and for writes of less than a block you have to read the existing 
 contents before the write to re-compute the parity.

   
Actually most (all?) raid 5 systems I've met don't check parity on read 
- they rely on the drive indicating a failed read.However the write 
splice penalty for raid 5 can still be pretty high (seek, read, insert 
new data, compute new parity, write data, seek, write parity) - and 
that's on top of the fact that the OS  did it's own logical read, check 
permissions, insert, write to update directories and the like.

John

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to connect to BackupPC server error

2007-03-27 Thread Winston Chan
On Mon, 2007-03-26 at 22:50 -0700, Craig Barratt wrote:
 Winston writes:
 
  I had been running BackupPC on an Ubuntu computer for several months to
  back the computer to a spare hard drive without problem. About the time
  I added a new host (Windows XP computer using Samba), I started getting
  the following behavior:
  
  BackupPC backs both hosts properly onto the spare hard drive once or
  twice after I reboot the Ubuntu server. Then I get a Error: Unable to
  connect to BackupPC server error when I attempt to go the web
  interface. When I restart BackupPC with /etc/init.d/backuppc restart,
  I get a message Can't create LOG file /var/lib/backuppc/log/LOG
  at /usr/share/backuppc/bin/BackupPC line 1735.
 
 Perhaps the /var/lib file system is full?
 
 If not, does the backuppc user have permissions to write in
 /var/lib/backuppc/log?
 
 Craig

I hadn't thought about the file system being full. After checking just
now, this is not the answer. /var/lib has 48G available on my main hard
drive. /var/lib/backuppc, to which the spare hard drive is mounted, has
59G available. 

The directory /var/lib/backuppc/log is owned by backuppc and is in the
backuppc group. Its permission is 750. All files in it also are owned by
backuppc and are in the backuppc group. They all have 640 permission.

Winston
 


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to connect to BackupPC server error

2007-03-27 Thread Winston Chan
It is a SATA drive as is the main drive.

Winston

On Tue, 2007-03-27 at 09:54 -0600, Ciarlotta, Aaron wrote:
 That spare hard drive wouldn't happen to be an external USB or Firewire,
 would it?
 
 
 Aaron Ciarlotta
 Linux/UNIX Systems Guru
 
  
 
 -Original Message-
 From: [EMAIL PROTECTED]
 [mailto:[EMAIL PROTECTED] On Behalf Of
 Winston Chan
 Sent: Monday, March 26, 2007 8:14 PM
 To: backuppc-users@lists.sourceforge.net
 Subject: [BackupPC-users] Unable to connect to BackupPC server error
 
 I had been running BackupPC on an Ubuntu computer for several months to
 back the computer to a spare hard drive without problem. About the time
 I added a new host (Windows XP computer using Samba), I started getting
 the following behavior:
 
 BackupPC backs both hosts properly onto the spare hard drive once or
 twice after I reboot the Ubuntu server. Then I get a Error: Unable to
 connect to BackupPC server error when I attempt to go the web
 interface. When I restart BackupPC with /etc/init.d/backuppc restart,
 I get a message Can't create LOG file /var/lib/backuppc/log/LOG at
 /usr/share/backuppc/bin/BackupPC line 1735.
 
 I have made sure that the backuppc is the owner of
 /var/lib/backuppc/log/LOG. It seems the only way to get backuppc to work
 again is to reboot the Ubuntu server. I then can see on the web
 interface that the last successful backup (of both hosts) occurred 1 or
 2 days after the previous reboot. Then backupPC works for 1 or 2 backups
 and the cycle starts again.
 
 What is the cause of this and how can I fix it?
 
 Winston
 
 
 
 
 -
 Take Surveys. Earn Cash. Influence the Future of IT Join
 SourceForge.net's Techsay panel and you'll get the chance to share your
 opinions on IT  business topics through brief surveys-and earn cash
 http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDE
 V
 ___
 BackupPC-users mailing list
 BackupPC-users@lists.sourceforge.net
 https://lists.sourceforge.net/lists/listinfo/backuppc-users
 http://backuppc.sourceforge.net/
 The information in this electronic mail message is sender's 
 business Confidential and may be legally privileged.  It is 
 intended solely for the addressee(s).  Access to this Internet 
 electronic mail message by anyone else is unauthorized.  If 
 you are not the intended recipient, any disclosure, copying, 
 distribution or any action taken or omitted to be taken in 
 reliance on it is prohibited and may be unlawful. The sender 
 believes that this E-mail and any attachments were free of 
 any virus, worm, Trojan horse, and/or malicious code when 
 sent. This message and its attachments could have been 
 infected during  transmission. By reading the message and 
 opening any attachments, the recipient accepts full 
 responsibility for taking protective and remedial action about 
 viruses and other defects. Travelport Inc. is not liable for any loss 
 or damage arising in any way from this message or its 
 attachments.
 


-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to connect to BackupPC server error

2007-03-27 Thread Les Mikesell
Winston Chan wrote:

 I had been running BackupPC on an Ubuntu computer for several months to
 back the computer to a spare hard drive without problem. About the time
 I added a new host (Windows XP computer using Samba), I started getting
 the following behavior:

 BackupPC backs both hosts properly onto the spare hard drive once or
 twice after I reboot the Ubuntu server. Then I get a Error: Unable to
 connect to BackupPC server error when I attempt to go the web
 interface. When I restart BackupPC with /etc/init.d/backuppc restart,
 I get a message Can't create LOG file /var/lib/backuppc/log/LOG
 at /usr/share/backuppc/bin/BackupPC line 1735.
 Perhaps the /var/lib file system is full?

 If not, does the backuppc user have permissions to write in
 /var/lib/backuppc/log?

 Craig
 
 I hadn't thought about the file system being full. After checking just
 now, this is not the answer. /var/lib has 48G available on my main hard
 drive. /var/lib/backuppc, to which the spare hard drive is mounted, has
 59G available. 
 
 The directory /var/lib/backuppc/log is owned by backuppc and is in the
 backuppc group. Its permission is 750. All files in it also are owned by
 backuppc and are in the backuppc group. They all have 640 permission.

There's also a possibility of running out of inodes.  Does
'df -i' look reasonable?

-- 
   Les Mikesell
[EMAIL PROTECTED]

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] RSync v. Tar

2007-03-27 Thread Jim McNamara

I have a bit of hard data to offer on this subject, as I recently switched a
backup from tar+ssh (over cygwin) to rsyncd.

The backuppc server is on the same physical LAN, and connect to each other
via a 192.168 address. All the cabling and switches support 100 MB full
duplex communications, and the servers have gigabit NICs. The backuppc
server is dumping the data to the /var partition which is on the 2 80Gb
satas in the case, running in a RAID 1 software array, under mdadm on Debian
stable.

These are the current stats, using rsyncd -

Backup#  Type  #Files  Size/MB  MB/sec  #Files  Size/MB  #Files
Size/MB  
30http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=30
full 168480
41758.5  5.66  168333  41639.8  310  118.9
35http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=35
incr 730
191.7  0.19  564  103.6  223  88.2
36http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=36
incr 785
198.4  0.18  725  123.0  106  75.4
37http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=37
full 169010
41836.6  5.73  168876  41750.2  276  86.5
38http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=38
incr 0
0.0  0.00  0  0.0  29  0.0
39http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=39
incr 0
0.0  0.00  0  0.0  0  0.0
40http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=40
incr 155
89.4  0.04  42  5.5  169  83.9
41http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=41
incr 321
124.2  0.05  142  19.8  234  104.4
Backup#  Type  Filled  Level  Start Date  Duration/mins  Age/days  Server
Backup Path  
30http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=30
full yes 0 3/16 18:00
122.9  11.2  /var/lib/backuppc/pc/sarah/30
35http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=35
incr no 1 3/21 18:00
16.9  6.2  /var/lib/backuppc/pc/sarah/35
36http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=36
incr no 1 3/22 18:00
17.9  5.2  /var/lib/backuppc/pc/sarah/36
37http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=37
full yes 0 3/23 18:00
121.7  4.2  /var/lib/backuppc/pc/sarah/37
38http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=38
incr no 1 3/24 18:00
16.5  3.2  /var/lib/backuppc/pc/sarah/38
39http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=39
incr no 1 3/25 18:00
15.4  2.2  /var/lib/backuppc/pc/sarah/39
40http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=40
incr no 1 3/26 18:00
33.1  1.2  /var/lib/backuppc/pc/sarah/40
41http://mail.stephanco.com/cgi-bin/BackupPC_Admin?action=browsehost=sarahnum=41
incr no 1 3/27 18:00
38.6  0.2
When it was doing tar, the full backups took far longer, in the neighborhood
of 600 minutes. The incremental backups took around an hour most days. So I
clearly made out better with rsyncd. Just as additional info, the server
being backed up is a file server for a small company. It is backing up the
directory where they store .jpg images of the products they sell. They
organize it by date, so obviously everything in the current day's directory
is new, but previous directories aren't modified most of the time.

I hope that helps.

Peace,
Jim

On 3/27/07, Holger Parplies [EMAIL PROTECTED] wrote:


Hi,

Les Mikesell wrote on 27.03.2007 at 01:03:32 [Re: [BackupPC-users] RSync
v. Tar]:
 Jesse Proudman wrote:
  I've got one customer who's server has taken 3600 minutes to
  backup.   77 Gigs of Data.  1,972,859 small files.  Would tar be
  better or make this faster?  It's directly connected via 100 Mbit to
 ^^^
  the backup box.
^^
 If the files don't change frequently, tar incremental runs will be much
 faster because they are based only on the target timestamps while rsync
will load the entire directory at both ends and compare them.

if you ask me, regardless of how your data changes, tar is the way to go,
not rsync, especially for *full* backups. With a direct 100 MBit
connection,
there's not much point in spending (lots of) CPU time for saving bandwidth
- not with 2 million files. rsync is good for low bandwidth connections,
where the link severely limits the transfer and speeding it up makes a
real
difference. In your case, your link speed is in the same order of
magnitude
as your disk I/O performance (considering our favorite topic, the seek
times
on the pool file system, the network link may in fact not even be the
limiting
factor - it clearly isn't, as 77 GB would take slightly more than 2 hours
to
transfer over a 100 MBit link, and rsync is not making it go faster than
that ;-).

rsync has additional benefits concerning finding and backing up new (or
moved) files with old timestamps and deleted files on *incremental*
backups,
but keeping the list of 

Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread nilesh vaghela

I found a strange characteristic of IDE drive.

When I connected CDROM and IDE hdd on same channel the data transfer with
hdparm -t command is @37 MB.

As I removed CDROM from that channel now the data transfer rate is @60 MB.

This seems to work with 4 to 5 linux pc.

--
Nilesh Vaghela
ElectroMech
Redhat Channel Partner and Training Partner
74, Nalanda Complex, Satellite Rd, Ahmedabad
25, The Emperor, Fatehgunj, Baroda.
www.electromech.info
-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] Unable to connect to BackupPC server error

2007-03-27 Thread Craig Barratt
Winston writes:

 I hadn't thought about the file system being full. After checking just
 now, this is not the answer. /var/lib has 48G available on my main hard
 drive. /var/lib/backuppc, to which the spare hard drive is mounted, has
 59G available. 
 
 The directory /var/lib/backuppc/log is owned by backuppc and is in the
 backuppc group. Its permission is 750. All files in it also are owned by
 backuppc and are in the backuppc group. They all have 640 permission.

Is that file system out of inodes?  Use df -i /var/lib/backuppc
to check.

Craig

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
Les Mikesell wrote:
 Evren Yurtesen wrote:
 
 #  Type  #Files  Size/MB  MB/sec  #Files  Size/MB  #Files Size/MB
 245 full  152228  2095.2   0.06   152177  2076.910818.3
 
 On Linux with raid setup with async io etc. people are getting 
 slightly better results. I think ufs2 is just fine. I wonder if there 
 is something in my explanations...The problem is backuppc. People are 
 getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives, 
 using Linux. It is a miracle that backup even finishes in 24 hours 
 using a standart ide drive.
 
 2MB/sec isn't bad when handling a lot of files (try unpacking a tar with 
  hundreds of thousands of little files to see).  The problem is that you 

Yes it is terrible. I get much better performance if I do the tar option 
with the same files. As a matter of fact I was using a smaller script 
for taking backups earlier. (which I still use on some servers) and 
transfer files over NFS. It works way faster, especially incremental 
backups take 5-10 minutes compared to 400 minutes with backuppc

 are getting .06 on a drive that is capable of running a couple of 
 sessions at 1.5MB/sec or so. I'd try to get up to normal speed before 
 claiming that the software needs to be fixed to get more.  More RAM will 
 probably make a huge difference, but first are you sure that your IDE 
 controller and cable are a match for the drive and that your OS is using 
 an efficient DMA mode?  I've seen even some of the 80-pin cables have a 
 problem that would make things shift down to 33Mhz pio mode which will 
 kill you.  Does freebsd log the IDE mode detected and have some way to 
 test throughput?

ad2: 238475MB ST3250824A/3.AAH [484521/16/63] at ata1-master UDMA100
DMA is working just fine. You are omitting the fact that 2mb/sec is very 
bad for a raid setup.

I agree, handling a lot of files might be slow but this depends on how 
you handle the files. But I was handling the same files before and it 
wasnt taking this long.

 BTW, how does BackupPC calculate speed? I think it calculates backup
 speed by reporting files transferred over time, so if you don't have
 many files that change, won't BackupPC report a very low backup speed.

 This is like the 'Contact' movie. The sphere took 30 seconds to 
 download but there were 18 hours of recording. If what you said was 
 true and backuppc would be backing up very small amount of files and 
 skipping most, then backups would probably take less time than 2-4 
 hours each.
 
 If you look at the 'duration' in the backup summary and the Size/MB in 
 the lower  File Size summary you can compute your own rate based on what 
 you are backing up.  For fulls at least this seems to be based on the 
 real target size, not what rsync had to transfer.
 

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
David Rees wrote:
 On 3/27/07, Evren Yurtesen [EMAIL PROTECTED] wrote:
 David Rees wrote:
 Evren, I didn't see that you mentioned a wall clock time for your
 backups? I want to know how many files are in a single backup, how
 much data is in that backup and how long it takes to perform that
 backup.
 I sent the status of the backups earlier today to mailing list?
 
 Still missing wall-clock time. Though we can extrapolate that it may
 be taking over 6 hours for a full backup of the system below. Is that
 correct?

That is true, full backups take about 500-600 minutes and incrementals 
take 200-300minutes.

 This is from 1 machine.

  Totals  Existing Files  New Files
 Backup# Type#Files  Size/MB MB/sec  #Files  Size/MB  
#Files  Size/MB
 112 full149290  1957.8  0.10149251  1937.9  460120.7
 
 I dont know if the problem is hard links. This is not a FreeBSD or Linux
 problem. It exists on both. Just that people using ultra fast 5 disk
 raid 5 setups are seeing 2mbytes/sec transfer rate means that backuppc
 is very very inefficient.
 
 As mentioned earlier, RAID 5 is horrible for random small read/write
 performance. It is often worse than a single disk.

Parity information is not required for normal reads, so you get 4-5 
times better performance with 5 disks. Plus usually those drives are 10K 
rpm. When I am making incremental or even full backups, backuppc does 
not write same files again so writes are very little. So you get a huge 
performance boost with raid5 and 10k rpm drives compared to a normal ide 
drive.

 But still, I have a client which has 1.5 million files and 80-100GB of
 data. A full backup takes 4-6 hours which is reasonable. Full backups
 average 4.5-5MB/s.
 
 For example this guy is using Linux (problem is OS independent)
 http://forum.psoft.net/showpost.php?p=107808postcount=16
 
 Your transfer rates are at least 5x slower than his and 10x slower
 than what most people here are getting. There is something wrong
 specifically with your setup. I suspect that if your backups were
 going 5x faster you'd be happy and 10x faster you'd be very happy.

I would be happy if backups were made with the same speed as other 
backup programs can do.

 On Linux with raid setup with async io etc. people are getting slightly
 better results. I think ufs2 is just fine. I wonder if there is
 something in my explanations...The problem is backuppc. People are
 getting ~2mbytes/sec(was it 2 or 5?) speed with raid5 and 5 drives,
 using Linux. It is a miracle that backup even finishes in 24 hours using
 a standart ide drive.
 
 With ext2 the default is async IO. With ext3 (the default system now)
 the default is ordered which is similar to BSD's soft updates.
 
 This is like the 'Contact' movie. The sphere took 30 seconds to download
 but there were 18 hours of recording. If what you said was true and
 backuppc would be backing up very small amount of files and skipping
 most, then backups would probably take less time than 2-4 hours each.
 
 Using the rsync transfer method, it does require at least a stat of
 every single file to be backed up. For 150k files and 2GB of data,
 you'd really expect this to be done within a hour with nearly any
 machine.

I am using rsync to sync 2GB of files from one machine to another and 
the number of files is about 100k and the operation takes 30seconds to 1 
minute in average.

 Are you using the --checksum-seed option? That will significantly
 speed up (the 3rd and later) backups as well.

No, it requires a patched rsync.

 Are you also sure the backup server isn't swapping as well? Can you
 get us some good logs from vmstat or similar during a backup?

I can tell that it is not swapping because the disk where the swaps are 
idle while taking backup. ad0 is the disk with the swap. If there was 
such problem then it would be reading like crazy from swap. I have given 
this information before.

 I also suspect that there is something else in this case slowing your
 machine down. Unless you give us the information to help track it
 down, we will not be able to figure it out. I feel like I am pulling
 teeth.
 

I have given you all the information you asked for(didnt I?), even tried 
async option. Incremental backup of 1 machine still took about 300 minutes.

The machine is working fine. I was using another backup program which 
was working way faster to backup the same machines. So I dont think that 
I have a hardware problem or such.

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net

Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Evren Yurtesen
nilesh vaghela wrote:
 I found a strange characteristic of IDE drive.
 
 When I connected CDROM and IDE hdd on same channel the data transfer 
 with hdparm -t command is @37 MB.
 
 As I removed CDROM from that channel now the data transfer rate is @60 MB.
 
 This seems to work with 4 to 5 linux pc.
 

I have no CDROMS installed on the machine and drives are on their own 
cables.

Thanks,
Evren

-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/


Re: [BackupPC-users] very slow backup speed

2007-03-27 Thread Adam Goryachev

 I have given you all the information you asked for(didnt I?), even tried 
 async option. Incremental backup of 1 machine still took about 300 minutes.

 The machine is working fine. I was using another backup program which 
 was working way faster to backup the same machines. So I dont think that 
 I have a hardware problem or such.
   
I haven't read *every* email in this thread (it was getting quite 
repetitive seeing you complaining about how crap backuppc is), but 
anyway, could you please elaborate on the method that your 'other' 
backup solution used, ie, if your other solution used rsync, and you are 
using rsync with backuppc, then that might be helpful to know. If you 
used tar before, and now use rsync, that would obviously make a difference.

I think overall, the attitude you have displayed is not conducive to 
resolving the problem. I also think it has been clearly shown that 
backuppc itself is not the problem, as other people are using it 
successfully, under similar conditions. What you need to do is basic 
diagnosis to find where the bottle neck is in your particular system.

I'd suggest you use whatever tools are available under your chosen OS (I 
think it is freebsd), to measure:
1) CPU utilisation on both the client being backed up, and the backuppc 
server
2) Memory/swap utilisation on both client and server
3) Disk utilisation/bandwidth on both client and server
4) Network utilisation/bandwidth on both client and server

Also, if you are using tar as your transfer method, then I'd suggest you 
try a 'manual' backup something like this (note, this command line is 
wrong, but you will need to adjust to get the right one).

from the backuppc machine:
cd /var/lib/backuppc;mkdir temp;cd temp;ssh client tar -cf - /backupdir 
| tar -xf -
Try and copy the command line args from your backup log file

Time how long that takes, and see how it compares to backuppc times 
if they are similar, then you know the issue is outside of backuppc.

Of course, if you are using rsync, then do the same test with rsync 
data, but I'm sure it has already been said that you should use tar for 
your case with lots of small files.

Just my 0.02c worth

Regards,
Adam

-- 
Adam Goryachev
Website Managers
Ph: +61 2 8304 [EMAIL PROTECTED]
Fax: +61 2 8304 0001www.websitemanagers.com.au



-
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT  business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.phpp=sourceforgeCID=DEVDEV
___
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/backuppc-users
http://backuppc.sourceforge.net/