Re: [Bacula-users] multiple spool files per job

2011-10-13 Thread Stefan Lubitz
Hi,

this feature would be really great. We have three Backup serves which all have 
single Jobs to backup (each job app. 13TB).
The Time to despool is absolutely waste of time and could be used more 
efficient. This feature request is more than 4 Years old and still not 
implemented.
Maybe I am wrong and I am sorry to say this, but as the data volume is growing 
all the time and this feature is still missing, I've got a notion that Bacula 
is not anymore up to date or state of the art.

Regards,
Stefan

-Ursprüngliche Nachricht-
Von: Alan Brown [mailto:a...@mssl.ucl.ac.uk] 
Gesendet: Montag, 10. Oktober 2011 13:00
An: James Harper
Cc: bacula-users@lists.sourceforge.net
Betreff: Re: [Bacula-users] multiple spool files per job

James Harper wrote:
 Is there a way to make bacula write multiple spool files per job? Two 
 would do. What I'm seeing is that 4 jobs start, all hit their spool 
 limit around the same time, then all wait in a queue until the file is 
 despooled. The despool happens fairly quickly (much quicker than the 
 spooling due to network and server fd throughput) so it isn't a huge 
 problem, but it would be better if the sd could just switch over to 
 another spool file when despooling starts so that the backup can 
 continue uninterrupted.

I believe this is in the todo list, but it will happen faster if someone 
submits already-written code.




--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2dcopy1
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users

--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2d-oct
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-13 Thread Thomas Mueller

On 13.10.2011 11:51, Stefan Lubitz wrote:
 Hi,

 this feature would be really great. We have three Backup serves which all 
 have single Jobs to backup (each job app. 13TB).
 The Time to despool is absolutely waste of time and could be used more 
 efficient. This feature request is more than 4 Years old and still not 
 implemented.
 Maybe I am wrong and I am sorry to say this, but as the data volume is 
 growing all the time and this feature is still missing, I've got a notion 
 that Bacula is not anymore up to date or state of the art.


as with any opensource tool you are free to send in code which enables 
the feature.

or hire somebody to write it for you and the community.

or buy the bacula enterprise version and tell them you really need the 
concurrent spooling/despooling feature.

- Thomas

 James Harper wrote:
 Is there a way to make bacula write multiple spool files per job? Two
 would do. What I'm seeing is that 4 jobs start, all hit their spool
 limit around the same time, then all wait in a queue until the file is
 despooled. The despool happens fairly quickly (much quicker than the
 spooling due to network and server fd throughput) so it isn't a huge
 problem, but it would be better if the sd could just switch over to
 another spool file when despooling starts so that the backup can
 continue uninterrupted.

 I believe this is in the todo list, but it will happen faster if someone 
 submits already-written code.




--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2d-oct
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-13 Thread Marcello Romani
Il 13/10/2011 11:51, Stefan Lubitz ha scritto:
 Hi,

 this feature would be really great. We have three Backup serves which all 
 have single Jobs to backup (each job app. 13TB).
 The Time to despool is absolutely waste of time and could be used more 
 efficient. This feature request is more than 4 Years old and still not 
 implemented.
 Maybe I am wrong and I am sorry to say this, but as the data volume is 
 growing all the time and this feature is still missing, I've got a notion 
 that Bacula is not anymore up to date or state of the art.

 Regards,
 Stefan


Hmmm... sounds like the perfect candidate for a paid enterprise feature 
request. I belive the creators of Bacula would be happy to hear from you.

Just a hint...


-- 
Marcello Romani

--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2d-oct
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-13 Thread Alan Brown
Marcello Romani wrote:

 Hmmm... sounds like the perfect candidate for a paid enterprise feature 
 request. I belive the creators of Bacula would be happy to hear from you.

I looked into this earlier this year.

Bacula Enterprise subscriptions are several _thousand_ euro per year, 
which is simply too much for us (and I daresay most academic sites).

I'm not sure why Kern has priced this out of the reach of most 
organisations who'd be willing to pay 500-900 euro/year.




--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2d-oct
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-13 Thread Marcello Romani
Il 13/10/2011 15:16, Alan Brown ha scritto:
 Marcello Romani wrote:

 Hmmm... sounds like the perfect candidate for a paid enterprise
 feature request. I belive the creators of Bacula would be happy to
 hear from you.

 I looked into this earlier this year.

 Bacula Enterprise subscriptions are several _thousand_ euro per year,
 which is simply too much for us (and I daresay most academic sites).

 I'm not sure why Kern has priced this out of the reach of most
 organisations who'd be willing to pay 500-900 euro/year.




I've never looked closely at their pricing, since we're a small shop and 
are doing reasonably well on our own so far, but if the prices are 
all-or-nothing in the range of thousands of dollars per year, as you 
write, then yes, I agree with you there's a gap in their support offering.

Or maybe there are other companies offering bacula support / consulting 
which are able to fill this gap (as I'd expect with an opensource 
enterprise-grade software)...

(ok I'll stop here since I've got nothing more to add than my own 
speculations :-)

-- 
Marcello Romani

--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2d-oct
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-10 Thread Alan Brown
James Harper wrote:
 Is there a way to make bacula write multiple spool files per job? Two
 would do. What I'm seeing is that 4 jobs start, all hit their spool
 limit around the same time, then all wait in a queue until the file is
 despooled. The despool happens fairly quickly (much quicker than the
 spooling due to network and server fd throughput) so it isn't a huge
 problem, but it would be better if the sd could just switch over to
 another spool file when despooling starts so that the backup can
 continue uninterrupted.

I believe this is in the todo list, but it will happen faster if someone 
submits already-written code.




--
All the data continuously generated in your IT infrastructure contains a
definitive record of customers, application performance, security
threats, fraudulent activity and more. Splunk takes this data and makes
sense of it. Business sense. IT sense. Common sense.
http://p.sf.net/sfu/splunk-d2dcopy1
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


[Bacula-users] multiple spool files per job

2011-10-08 Thread James Harper
Is there a way to make bacula write multiple spool files per job? Two
would do. What I'm seeing is that 4 jobs start, all hit their spool
limit around the same time, then all wait in a queue until the file is
despooled. The despool happens fairly quickly (much quicker than the
spooling due to network and server fd throughput) so it isn't a huge
problem, but it would be better if the sd could just switch over to
another spool file when despooling starts so that the backup can
continue uninterrupted.

I'm spooling to internal RAID, then despooling to external USB. While
spooling isn't really advised when the backup target is a disk, doing it
this way means I can run multiple jobs at once without causing
interleaving in the backup file (single sd volume) or severe filesystem
fragmentation (if one sd volume per job). Internal RAID writes at
~100MB/second while the USB disk writes at ~30MB/second so it turns out
to be a pretty effective way to do what I want except that despooling is
causing a bottleneck.

Any suggestions?

Thanks

James



--
All of the data generated in your IT infrastructure is seriously valuable.
Why? It contains a definitive record of application performance, security
threats, fraudulent activity, and more. Splunk takes this data and makes
sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-d2dcopy2
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users


Re: [Bacula-users] multiple spool files per job

2011-10-08 Thread Ralf Gross
James Harper schrieb:
 Is there a way to make bacula write multiple spool files per job? Two
 would do. What I'm seeing is that 4 jobs start, all hit their spool
 limit around the same time, then all wait in a queue until the file is
 despooled. The despool happens fairly quickly (much quicker than the
 spooling due to network and server fd throughput) so it isn't a huge
 problem, but it would be better if the sd could just switch over to
 another spool file when despooling starts so that the backup can
 continue uninterrupted.
 
 I'm spooling to internal RAID, then despooling to external USB. While
 spooling isn't really advised when the backup target is a disk, doing it
 this way means I can run multiple jobs at once without causing
 interleaving in the backup file (single sd volume) or severe filesystem
 fragmentation (if one sd volume per job). Internal RAID writes at
 ~100MB/second while the USB disk writes at ~30MB/second so it turns out
 to be a pretty effective way to do what I want except that despooling is
 causing a bottleneck.
 
 Any suggestions?

No, this has been on the feature requests list for a while now.
Spooling nearly doubles my time for large backups. 

http://thread.gmane.org/gmane.comp.sysutils.backup.bacula.devel/15351

Item 10: Concurrent spooling and despooling within a single job.
http://www.bacula.org/git/cgit.cgi/bacula/plain/bacula/projects?h=Branch-5.1

Ralf

--
All of the data generated in your IT infrastructure is seriously valuable.
Why? It contains a definitive record of application performance, security
threats, fraudulent activity, and more. Splunk takes this data and makes
sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-d2dcopy2
___
Bacula-users mailing list
Bacula-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/bacula-users