Tim,

First off, thank you very much for your insightful and helpful approach, 
despite the limited information I provided. And thank you as well for being the 
“unpaid” mechanic (good analogy). I will try to present a detail account. 
Please read my comments to your questions below:

Environment:

·         Server (using gigabit wired Ethernet on a managed switch)

o   running under VMWare ESX5.5

o   VM allocated w/ 4gb ram, all procs/cores, and max cpu/mem bursts

o   intel i5-3330 3GHz

o   24gb ram

o   Backup location 5 disk hardware raid 5 under Centos 6.5

§  Backup avail space 6 TB

·         Windows Workstation Client (using gigabit wired Ethernet on a managed 
switch)

o   Intel i7-4770K cpu 3.5GHz

o   32 gb ram

o   2 internal logical drives

§  C: HW raid 0 comprising of TWO 120GB SSD drives (238gb capacity)

·         89.5GB used (162,337 files)

§  D: HW raid 5 consisting of 4 x 2TB (5589gb capacity)

·         1.5TB used (213,594 files)

I have not completed a full backup of this workstation yet. I have done many 
incremental removing exclusions in an attempt to incrementally backup the 
entire computer (drive c &d) in steps.
I will disable compression and re-attempt, although I will not be able to 
report the information for xx days until its complete.


> Now, these are established backups.  I've found that initial backups can be 
> 2-3 times as long.  And at 3TB, you're nearly triple of the bigger one.  So, 
> that could mean that the initial backup can take 6000
> minutes or more, or four days.  And personally, I've found that compression 
> can double or triple the time again.  Worst-case of all of that could be two 
> weeks!  So it is possible that you're seeing correct
> operation for that first backup, and that future backups should be much 
> better.
>
So I guess compression is out. I will disable. Thank you for the info.


> Also, those are fulls.  Incrementals take 4 hours and 1 hour, respectively.
>
> Is rsync “really” that slow?
>
> No:  rsync is only going to be marginally slower than a non-rsync copy, even 
> on the first time, assuming you're not limited by something else (CPU or RAM) 
> that would not be a limit for a normal copy.
>
> That could be related to the number of files:  that's an area where rsync can 
> get tripped up.  As you can see, I've got >1 million files, so the definition 
> of "too many" is pretty big.  But if you had, say, 10M files,
> maybe that's an issue to consider.

I don’t have anywhere near a million files, so I guess my workstation should be 
quicker. I suspect vmware is causing *some* penalty, but it should not be that 
great. Regardless, it’s what I have so that’s got to work fine.


> Yes.  First, you've given us *NOTHING* to go on other than "it's slow".  It's 
> like telling a mechanic "my car doesn't run right."  Of course, you're 
> probably expecting to pay the mechanic, so he's incented to ask
> lots of questions to figure things out.  I think it's pretty telling that I 
> can't see a single reply to your request.  While your request includes a lot 
> of words, there was almost *NOTHING* of substance in your
> request except a *VERY* brief description of your hardware.  We have almost 
> no description of what your backups look like (size and number of files, for 
> example), what your backup jobs look like (compression
> being a very big one), or what your backup server is doing (CPU, memory or 
> disk usage).   And we're not getting paid, so we're not really incented to 
> ask you a lot of questions.  But I'll give you a few:
>

Yes, I’m sorry. Your points are well made. I’m trying to correct my mistake in 
this email.

> First, is your system limited in some way?  Are you hitting 100% CPU, or 
> swapping (not enough RAM), or are your disk overloaded, or something else?  
> Learn to use top and vmstat (or dstat).

CPU normally sits around 50%-70% depending if a 2nd job kicks in. I’ve used 
“iostat -k 5” and %iowait from 0% to as much as 40% but usualy hovers around 
10%-20%. I was unaware of dstat, but I’ll start running it when I perform the 
full backup

While doing some googling, I read somewhere about adjusting the TCP buffers for 
rsync. I’ve applied the following line in my rsync.conf file on the windows 
machine:
socket options = TCP_NODELAY SO_RCVBUF=8192 SO_SNDBUF=65536
Have you experimented with this or something similar in order to improve 
performance? I don’t have any hard numbers with/without (yet), but thought I 
would mention this.

> Is your infrastructure fundamentally limited in some way?  Have you tried 
> doing a straight-up copy from a target to your backup system to make sure 
> that the underlying infrastructure is capable of delivering
> what you expect it to?  If you can only get 1-2MB/s copies using SMB, tar, 
> NFS, FTP, etc. , then that's all you'll get with BackupPC, too.  But if you 
> can get 70MB/s copies between the same two systems some
> other way, then we can expect better of BackupPC.  (But all that does is 
> re-ask the question of what is limiting you.)

No, my baseline was based on my prior experience with bacula, and a trial of 
Retrospect (which I quickly discounted). Both software I was able to fully back 
up this workstation within 4-8 hrs—not days. This is what started to get me 
worried.

> From my e-mail, you know what is possible and reasonable to get.  If you're 
> far away from those results, then you need to figure out what is different 
> about your system and causing the slowdown.

> The second thing to try is to simplify things.  For me, the first thing I do 
> is disable compression.  In today's multi-core universe, compression is 
> rapidly becoming a bottleneck again.  The compression algorithms
> in common use today do *not* use multiple cores.  On a system with more than 
> a couple of disks I can easily max out one core with compression.
Yes, I learned this while experimenting between gzip and pigz. For now, I’ll 
disable compression.

> Another option would be to expand your testing from a very small section to 
> something larger:  say, 100GB.  That is big enough to be somewhat 
> representative of the whole, but should be able to complete
> quickly enough even with compression, encryption, etc. to get some baseline 
> numbers to work with, including both the *first* and *additional* full 
> backups.  That way, you might find that the initial backup will
> take a week, but each additional backup after that will only take 12 hours 
> and you're OK with that.  Or, you might find that things are still broken, 
> but now it won't cost you a week of your life every time you
> want to test.
GREAT idea. Before I embark on trying to do a full backup, I’ll isolate ONE 
folder with some large files (around 100gb) and focus on doing some tests on 
that with/without compression. I will report my findings.

> If it helps, the server is beefy, quad-core i5 proc w/ 4gb ram and 6TB raid5.

Tim, again, thank you for taking the time and interest to both help and point 
out the vagueness of the email. I will get back with you.

Thanks,

Marco

------------------------------------------------------------------------------
"Accelerate Dev Cycles with Automated Cross-Browser Testing - For FREE
Instantly run your Selenium tests across 300+ browser/OS combos.
Get unparalleled scalability from the best Selenium testing platform available
Simple to use. Nothing to install. Get started now for free."
http://p.sf.net/sfu/SauceLabs
_______________________________________________
BackupPC-users mailing list
BackupPC-users@lists.sourceforge.net
List:    https://lists.sourceforge.net/lists/listinfo/backuppc-users
Wiki:    http://backuppc.wiki.sourceforge.net
Project: http://backuppc.sourceforge.net/

Reply via email to