Hi Brock

Thanks for the quick response 

The filling up of diskspace seem to be the cause of the failing of my jobs. 
I will apply compression to save on disk space. 

I have a question about globs you mention. I can't say I know what that is. 
How would you apply this in the a config file?

The large volume I picked is the home folder for users. So i thought of 
creating a jobdef that includes certain number of folders and a second and 
even a third. Then creating jobs for each of the separately.

What is your opinion about this?

Regards
Waanie



On Thursday, June 25, 2020 at 4:31:15 PM UTC+2, Brock Palen wrote:
>
> If you are using disk volumes, you probably want to auto label to get new 
> volumes created as needed. 
> Note bareos tries to preserve as much data as possilbe and with disk 
> volumes I find it likes to fill the disk and eventually fail. 
>
> I run an ‘admin’ job that just runs a script on a schedule, you could use 
> cron etc that checks volumes so they correctly get pruned and then 
> truncated (made zero size) to free disk space, you will need to modify for 
> your pool names: 
>
> #!/bin/bash 
> for x in `echo "list volumes pool=AI-Consolidated" | bconsole | grep -v 
> "list volumes" | grep AI-Consolid | awk -F\| '{print $3}'` 
>  do 
>  echo "prune volume=$x yes" 
> done | bconsole 
>
> # actaully free up disk space 
> echo "truncate volstatus=Purged pool=AI-Consolidated yes" \ 
> | bconsole 
>
>
> As for the very large backup few ideas 
>
> * use globs and break it into multiple jobs (this won’t impact restores) 
> * number of files will dictate scan time for incremental rather than size 
> of data. 
>     Test scan time with estimate:   estimate job=<jobname>  accurate=yes 
> level=Incremental 
> * Fulls are dominated by bandwidth 
>   ** Compression will cause CPU to peak. and limit performanceif not 
> IO/Network bound 
>   ** If using compression look at the low cpu/compression trade off 
> options 
>   ** Maybe not compress your backup but use a migrate job with a compress 
> filter to compress all on the backup server 
> * fd compression is single threaded, if you break it into multiple jobs 
> with globs you can run multiple at a time 
>
> You're going to want to benchmark all along your system,  I like dstat 
> over iostat/top etc for monitoring. but a 90TB single volume backup will 
> take some time for a full.  If you have the space on your server maybe look 
> at Always Incremental,  so you never actually make a full download of that 
> volume again, though you will copy 90TByte of data on your SD every few 
> days depending on settings, just like multiple fulls. 
>
> Myself I have ‘archive’  jobs where I take VirtualFull of each of my jobs 
> in a different volume on a different media monthly for safety.  Bailed me 
> out a lot when I was learning. 
>
> Brock Palen 
> 1 (989) 277-6075 
> [email protected] <javascript:> 
> www.mlds-networks.com 
> Websites, Linux, Hosting, Joomla, Consulting 
>
>
>
> > On Jun 24, 2020, at 4:05 PM, Waanie Simon <[email protected] 
> <javascript:>> wrote: 
> > 
> > Hi all 
> > 
> > I am working on building a backup solution with bareos. the system will 
> be used to replace arkeia which is no longer supported. Arkeia backups were 
> quite easy to install and monitor but we are. 
> > 
> > I have install bareos-director and bareos-sd on the same server. They 
> are running on the same network interface. 
> > 
> > My initial backups went fine but as time went on the backups would 
> queued. Including the catalog backup. 
> > I sometime see that the system is asking for a labeling to happen but it 
> doesn't always fix the problem. 
> > 
> > mostly linux vms and proxmox physical servers. 
> > 
> > Some of the harder ones to backup is our file server with home volume of 
> 90 TB. This is something that we could never backup and has become a source 
> of many problems. 
> > 
> > How would you backup such a volume. Should it be split? Since there is a 
> lot of static content in there, should it be archived? 
> > 
> > I am not sure if config files are is set correctly. 
> > 
> > I have incremental backups happening daily and differentials happening 
> weekly but I often have full backups happening multiple times during a 
> month. 
> > 
> > Currently we have no tape library in place so all is running on disk. 
> > 
> > Thought I could add some code snippets to give an idea of what I have 
> going on here 
> > 
> > 
> > client 
> > 
> > 
> > Client { 
> >   Name = dr001-fd 
> >   Address = 10.60.100.12 
> >   Password = <passwd> 
> > } 
> > 
> > Job 
> > 
> > Job { 
> >   Name = dr001-Daily-inc 
> >   JobDefs = linux-daily-inc 
> >   Type = backup 
> >   Messages = Standard 
> >   Client = dr001-fd 
> > } 
> > 
> > 
> > JobDefs { 
> >   Name = linux-daily-inc 
> >   Type = Backup 
> >   Level = Incremental 
> >   Storage = File 
> >   Pool = Incremental 
> >   FileSet = LinuxAll 
> > } 
> > 
> > each job has its own schedule. is this necessary 
> > 
> > Schedule { 
> >   Name = dr001-daily-inc 
> >   Description = Incremental Daily for dr001-fd 
> >   Run = daily at 11:00 
> > } 
> > 
> > 
> > storage under the bareos-dir.d folder 
> > File.conf file 
> > 
> > Storage { 
> >   Name = File 
> >   Address = ctbackup.cape.saao.ac.za                # N.B. Use a fully 
> qualified name here (do not use "localhost" here). 
> >   Password = <password> 
> >   Device = FileStorage 
> >   Media Type = File 
> >   Maximum Concurrent Jobs = 5 
> > } 
> > 
> > 
> > The Pools are configured as 
> > 
> > Pool { 
> >   Name = Incremental 
> >   Pool Type = Backup 
> >   Recycle = yes                       # Bareos can automatically recycle 
> Volumes 
> >   AutoPrune = yes                     # Prune expired volumes 
> >   Volume Retention = 30 days          # How long should the Incremental 
> Backups be kept?  (#12) 
> >   Maximum Volume Bytes = 150G           # Limit Volume size to something 
> reasonable 
> >   Maximum Volumes = 30               # Limit number of Volumes in Pool 
> >   Label Format = "Incremental-"       # Volumes will be labeled 
> "Incremental-<volume-id>" 
> > } 
> > 
> > Pool { 
> >   Name = Differential 
> >   Pool Type = Backup 
> >   Recycle = yes                       # Bareos can automatically recycle 
> Volumes 
> >   AutoPrune = yes                     # Prune expired volumes 
> >   Volume Retention = 90 days          # How long should the Differential 
> Backups be kept? (#09) 
> >   Maximum Volume Bytes = 100G          # Limit Volume size to something 
> reasonable 
> >   Maximum Volumes = 60               # Limit number of Volumes in Pool 
> >   Label Format = "Differential-"      # Volumes will be labeled 
> "Differential-<volume-id>" 
> > } 
> > 
> > Pool { 
> >   Name = Full 
> >   Pool Type = Backup 
> >   Recycle = yes                       # Bareos can automatically recycle 
> Volumes 
> >   AutoPrune = yes                     # Prune expired volumes 
> >   Volume Retention = 365 days         # How long should the Full Backups 
> be kept? (#06) 
> >   Maximum Volume Bytes = 350G          # Limit Volume size to something 
> reasonable 
> >   Maximum Volumes = 100               # Limit number of Volumes in Pool 
> >   Label Format = "Full-"              # Volumes will be labeled 
> "Full-<volume-id>" 
> > } 
> > 
> > 
> > unfortunately there has been a bit of a thumb suck regarding the numbers 
> here. 
> > 
> > The storage config looks like this 
> > 
> > 
> > Devices 
> > 
> > Device { 
> >   Name = FileStorage 
> >   Media Type = File 
> >   Archive Device = /data1/bareos/FileStorage 
> >   LabelMedia = yes;                   # lets Bareos label unlabeled 
> media 
> >   Random Access = yes; 
> >   AutomaticMount = yes;               # when device opened, read it 
> >   RemovableMedia = no; 
> >   Collect Statistics = yes 
> >   AlwaysOpen = no; 
> >   Description = "File device. A connecting Director must have the same 
> Name and MediaType." 
> > } 
> > 
> > 
> > bareos-sd.conf 
> > 
> > Storage { 
> >   Name = bareos-sd 
> >   Maximum Concurrent Jobs = 20 
> > 
> >   # remove comment from "Plugin Directory" to load plugins from 
> specified directory. 
> >   # if "Plugin Names" is defined, only the specified plugins will be 
> loaded, 
> >   # otherwise all storage plugins (*-sd.so) from the "Plugin Directory". 
> >   # 
> >   # Plugin Directory = "/usr/lib/bareos/plugins" 
> >   # Plugin Names = "" 
> >   Collect Device Statistics = yes 
> >   Collect Job Statistics = yes 
> >   #Statistics Collect Intevals = 60 
> > } 
> > ~ 
> > 
> > 
> > 
> > I know that my largest backups will probably not work since my capacity 
> I can write to is only about 70 Tb 
> > 
> > I have the web gui working but there is no easy way to see progress on a 
> job. 
> > 
> > Any improvements would be appreciated 
> > 
> > Regards 
> > Waanie 
> > ~   
> > 
> > -- 
> > You received this message because you are subscribed to the Google 
> Groups "bareos-users" group. 
> > To unsubscribe from this group and stop receiving emails from it, send 
> an email to [email protected] <javascript:>. 
> > To view this discussion on the web visit 
> https://groups.google.com/d/msgid/bareos-users/c8c2054a-3d75-4bad-9a97-f2674e08c4b0o%40googlegroups.com.
>  
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/eed2aa23-cf27-406c-944a-f1d6a57823c7o%40googlegroups.com.

Reply via email to