On Friday, July 10, 2015 at 7:43:45 AM UTC+2, Bruno Friedmann wrote:
> On Thursday 09 July 2015 18.13:16 Bruno Marsal wrote:
> > Hi, consider following setup:
> > * Storage daemon configured with archive device: /mnt/backup with media
> > type: file
> > * Storage daemon connected via 10Gbps Ethernet
> > * Many jobs. For example: Backup server1, Backup server2, . Backup server70
> > * Each server has its own volume pool: server1-fd-Pool, server2-fd-Pool, .
> > server70-df-Pool
> > 
> > This configuration worked great but the problem was the jobs are running 
> > consecutively and would run sometimes several days (in case of full 
> > backups). Seeking for an option to run the backup jobs in parallel I created
> > 4 Device definitions on the storage daemon and 4 storage definitions on the
> > director:
> > On Storage daemon:
> > * Device { Name = Backup_01; Archive device = /mnt/backup; media type = 
> > file; [...]}
> > * Device { Name = Backup_02; Archive device = /mnt/backup; media type = 
> > file; [...]}
> > * Device { Name = Backup_03; Archive device = /mnt/backup; media type = 
> > file; [...]}
> > * Device { Name = Backup_04; Archive device = /mnt/backup; media type = 
> > file; [...]} On Director:
> > * Storage { Name = Storage_01; Device = Backup_01; [...]} ...
> > * Storage { Name = Storage_04; Device = Backup_04; [...]}
> > 
> > The pools were randomly configured to use either Storage_01, Storage_02,
> > Storage_03 and Storage_04 --> Now 4 jobs are running in parallel which is 
> > almost perfect.
> > 
> > The problem with this configuration is: Whenever a new server is being 
> > added to be backed up, we manually must choose a storage (01-04) which is 
> > all but perfect.
> > Is there a way to configure/modify the job scheduler in order it assigns 
> > one storage after the other to the scheduled jobs? Could this be done using 
> > plugins?
> > Is there some other nicer way to more or less automatically balance the 
> > jobs to the available storages?
> > 
> > Bruno
> > 
> 
> What you're looking for is "spooling", so you could have on storage and 
> device, and all jobs will be spooled together
> then for x reasons write done to the device. 
> 
> 
> -- 
> 
> Bruno Friedmann 
> Ioda-Net Sàrl www.ioda-net.ch
>  
>  openSUSE Member & Board, fsfe fellowship
>  GPG KEY : D5C9B751C4653227
>  irc: tigerfoot

Thank you for the answer. Will consider using a single pool for all servers/job.

I hoped there is a way to comfortably manage a pool per server but still have 
them on a single fileserver and be able to run backups in parallel. Seems there 
is no easy way to modify the scheduler to fit my needs. At least none I am 
aware of.

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to