Re: [Bacula-users] Windows client performance

2010-01-25 Thread mehma sarja
Timo,.

Your test involved just 600 some MB of data which may not be a large or
varied enough data set. There is data and then there is data. Millions of
small email index files is harder on Bacula than thousands of large files.
And you don't mentioned if or how you controlled for your network bandwidth
- did you keep all traffic off it?

I'd suggest you use a data set which is typical of your intended use and
multiply it by a factor of 10 and keep bandwidth stable.

Mehma
===
--
Throughout its 18-year history, RSA Conference consistently attracts the
world's best and brightest in the field, creating opportunities for Conference
attendees to learn about information security's most important issues through
interactions with peers, luminaries and emerging and established companies.
http://p.sf.net/sfu/rsaconf-dev2dev___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Windows client performance

2010-01-25 Thread Mike Ruskai
On 1/25/2010 11:07 AM, Timo Neuvonen wrote:
 some more cpu load, but no peak in workstation's cpu load meter exceeded
 50%.


In Windows, 100% load means all CPU's together at max load.  If you have 
two cores, 50% means 100% load on one core.

So you're seeing the best that CPU can do (and P4's are a seriously 
awful design), short of a major change in the compression library to 
utilize multiple threads.


--
Throughout its 18-year history, RSA Conference consistently attracts the
world's best and brightest in the field, creating opportunities for Conference
attendees to learn about information security's most important issues through
interactions with peers, luminaries and emerging and established companies.
http://p.sf.net/sfu/rsaconf-dev2dev
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Windows client performance

2010-01-25 Thread Timo Neuvonen
Mike Ruskai than...@earthlink.net kirjoitti viestissä 
news:4b5dcf2f.9050...@earthlink.net...
 On 1/25/2010 11:07 AM, Timo Neuvonen wrote:
 some more cpu load, but no peak in workstation's cpu load meter exceeded
 50%.


 In Windows, 100% load means all CPU's together at max load.  If you have
 two cores, 50% means 100% load on one core.


Actually, there were two graphs, one per core. Neither of them peaked close 
to 50% at GZIP1 level. After increasing comprerssion level (up to GZIP9), it 
needed more cpu load, but it wasn't absolutely the limiting factor if the 
graphs are somehow close to truth.

But the basic question still is, what limits the data rate without 
compression, when cpu load is very low? Using compression still drops the 
compressed rate by same factor the data size shrinks, while the required 
real time doesn't practically chage at all. Sounds very much like it was a 
limit of disk io rate, but it sounds unbelievable to me since figures are 
that low.

But maybe I'll need to test disk read performance first by some other means. 
It will just take a couple of days until I'll have time for that project 
again.


 So you're seeing the best that CPU can do (and P4's are a seriously
 awful design), short of a major change in the compression library to
 utilize multiple threads.




--
The Planet: dedicated and managed hosting, cloud storage, colocation
Stay online with enterprise data centers and the best network in the business
Choose flexible plans and management services without long-term contracts
Personal 24x7 support from experience hosting pros just a phone call away.
http://p.sf.net/sfu/theplanet-com
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Windows client performance

2010-01-25 Thread Timo Neuvonen
mehma sarja mehmasa...@gmail.com kirjoitti viestissä
news:ec5d34681001250830h9cfb180naea17beb1a544...@mail.gmail.com...
 Timo,.

Your test involved just 600 some MB of data which may not be a large or
varied enough data set. There is data and then there is data. Millions of
small email index files is harder on Bacula than thousands of large files.
And you don't mentioned if or how you controlled for your network
bandwidth - did you keep all traffic off it?

 I'd suggest you use a data set which is typical of your intended use and
 multiply it by a factor of 10 and keep bandwidth stable.

Network load definetely wasn't the reason here, hard to document it, but I 
just know it. Also, the results were very repeatable when I run the same 
test several times, this doesn't sound like a network load problem either.

But yes, I'll need to re-run the test with another type of data. A lot of 
small files will cause plenty of overhead. I just didn't think it would be 
that much overhead...

--
Timo 



--
The Planet: dedicated and managed hosting, cloud storage, colocation
Stay online with enterprise data centers and the best network in the business
Choose flexible plans and management services without long-term contracts
Personal 24x7 support from experience hosting pros just a phone call away.
http://p.sf.net/sfu/theplanet-com
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Windows client performance

2010-01-25 Thread Mike Ruskai

On 1/25/2010 2:46 PM, Timo Neuvonen wrote:

Mike Ruskaithan...@earthlink.net  kirjoitti viestissä
news:4b5dcf2f.9050...@earthlink.net...
   

On 1/25/2010 11:07 AM, Timo Neuvonen wrote:
 

some more cpu load, but no peak in workstation's cpu load meter exceeded
50%.


   

In Windows, 100% load means all CPU's together at max load.  If you have
two cores, 50% means 100% load on one core.

 


Actually, there were two graphs, one per core. Neither of them peaked close
to 50% at GZIP1 level. After increasing comprerssion level (up to GZIP9), it
needed more cpu load, but it wasn't absolutely the limiting factor if the
graphs are somehow close to truth.

   
You can't go by Task Manager graphs, either.  Unless you set process 
affinity, the thread will be scheduled on alternating cores, and the 
graphs won't accurately represent how loaded each core is over time.  If 
you have a single thread running at full speed, it will very likely look 
like roughly 50% usage on each core over time.  The calculations just 
don't happen with a granularity small enough to show the actual usage.  
If you look at the CPU usage figure on the bottom left, however, that 
will be reasonably accurate.  If it reads 50%, then the compression work 
is going as fast as it can.

But the basic question still is, what limits the data rate without
compression, when cpu load is very low? Using compression still drops the
compressed rate by same factor the data size shrinks, while the required
real time doesn't practically chage at all. Sounds very much like it was a
limit of disk io rate, but it sounds unbelievable to me since figures are
that low.

But maybe I'll need to test disk read performance first by some other means.
It will just take a couple of days until I'll have time for that project
again.


   
It could be either disc I/O or network I/O that's bottlenecking.  The 
Bacula file daemon for Windows is definitely not restricted to the rates 
you're seeing.



--
The Planet: dedicated and managed hosting, cloud storage, colocation
Stay online with enterprise data centers and the best network in the business
Choose flexible plans and management services without long-term contracts
Personal 24x7 support from experience hosting pros just a phone call away.
http://p.sf.net/sfu/theplanet-com___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] Windows client performance

2010-01-25 Thread James Harper
   Elapsed time:   2 mins 35 secs
   Priority:   10
   FD Files Written:   4,648
   SD Files Written:   4,648
   FD Bytes Written:   664,739,011 (664.7 MB)
   SD Bytes Written:   665,458,318 (665.4 MB)
   Rate:   4288.6 KB/s
   Software Compression:   None
   VSS:yes
   Encryption: no
   Accurate:   no

How long does it take to copy those files to a share on another windows
computer? If you test that, make sure you 'pull' the files from another
computer rather than 'push' them (eg originate the copy on the other
computer). That should test your network and disk IO without involving
Bacula. It would have to be an order of magnitude different at least to
draw any real conclusions though.

 
 Then, after adding lowest-level (GZIP1) compression:
 
 Elapsed time:   2 mins 43 secs
   Priority:   10
   FD Files Written:   4,648
   SD Files Written:   4,648
   FD Bytes Written:   308,509,378 (308.5 MB)
   SD Bytes Written:   309,228,685 (309.2 MB)
   Rate:   1892.7 KB/s
   Software Compression:   53.6 %
   VSS:yes
   Encryption: no
   Accurate:   no
 

Can you zip up the files to another computer? That might be a bit
tricker to organise in a way that is similar to how bacula works...

 Any ideas what is the bottleneck here? I think 1GB of memory should be
 enough for XP alone. Some glue makes me think about the disk, but this
rate
 is a decade below what to expect from even that old hard disk.
 

The XP firewall at around sp2 was kind of broken in that the various tcp
offloads didn't work properly and slowed things down to a crawl. Try
disabling the firewall service in services.msc - just turning off the
windows firewall on the network adapter is not sufficient, you have to
stop the service.

James


--
The Planet: dedicated and managed hosting, cloud storage, colocation
Stay online with enterprise data centers and the best network in the business
Choose flexible plans and management services without long-term contracts
Personal 24x7 support from experience hosting pros just a phone call away.
http://p.sf.net/sfu/theplanet-com
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users