Hi All,
does anyone can share me some clue or insights on how I can take care this
kind of case?
I still cannot find which part that limits my running backups.
Any clue/insights will be much appreciated.
Thank you.
On Friday, October 29, 2021 at 10:13:23 AM UTC+7 Reqi Rangga wrote:
> Hi All,
>
> hope you guys can find this thread well.
> I still have this case as my blocker on my research about how bareos can
> perform on a large backup scale.
> Does anyone have same experience in the past? or you guys can share me
> some insights on what to look for?
>
> Thank you.
>
> On Tuesday, October 26, 2021 at 9:42:28 AM UTC+7 Reqi Rangga wrote:
>
>> Hi All,
>>
>> I was wondering whether any of you have any suggestion that might relate
>> to my use case?
>> I would like to provide more information in case you need more deeper
>> view.
>>
>> Thank you.
>>
>> Regards,
>> Reqi Rangga
>>
>> On Friday, October 22, 2021 at 9:46:31 AM UTC+7 Reqi Rangga wrote:
>>
>>> Hi All,
>>>
>>> is anyone have ever experience same issue? or have any insights that you
>>> guys can share with me?
>>>
>>> Thank you.
>>> Regards,
>>>
>>> Reqi Rangga
>>>
>>> On Tuesday, October 19, 2021 at 10:51:46 AM UTC+7 Reqi Rangga wrote:
>>>
>>>> Hi All,
>>>>
>>>> I tested Bareos in my testing environment to backup our VMs in our
>>>> vcenter environment and got issue when I'm trying to run approx 20++
>>>> concurrent job, the 20th job and the rest always in waiting status where
>>>> it
>>>> cannot find any appendable volumes and ask us to create new volume
>>>> manually
>>>> using 'label' command.
>>>>
>>>> This is my environment and config :
>>>> *Bareos Director*
>>>> *status director
>>>> temng02bosdir01-dir Version: 20.0.1 (02 March 2021) CentOS Linux
>>>> release 7.6.1810 (Core)
>>>>
>>>> *show director
>>>> Director {
>>>> Name = "temng02bosdir01-dir"
>>>> Messages = "Daemon"
>>>> DirAddresses = {
>>>> ipv4 = {
>>>> addr = 0.0.0.0
>>>> port = 9101
>>>> }
>>>> }
>>>> QueryFile = "/usr/lib/bareos/scripts/query.sql"
>>>> MaximumConcurrentJobs = 30
>>>> Password = "[md5]"
>>>> Auditing = Yes
>>>> }
>>>>
>>>> *show storage
>>>> Storage {
>>>> Name = "File"
>>>> Address = "temng02bossd01.biznetgiocloud.net"
>>>> Password = "[md5]"
>>>> Device = "FileStorage", "FileStorage1", "FileStorage2",
>>>> "FileStorage3", "FileStorage4", "FileStorage5", "FileStorage6",
>>>> "FileStorage7", "FileStorage8", "FileStorage9", "FileStorage10",
>>>> "FileStorage11", "FileStorage12", "FileStorage13", "FileStorage14",
>>>> "FileStorage15", "FileStorage16", "FileStorage17", "FileStorage18",
>>>> "FileStorage19", "FileStorage20", "FileStorage21", "FileStorage22",
>>>> "FileStorage23", "FileStorage24", "FileStorage25", "FileStorage26",
>>>> "FileStorage27", "FileStorage28", "FileStorage29"
>>>> MediaType = "File"
>>>> MaximumConcurrentJobs = 30
>>>> }
>>>>
>>>> /etc/bareos/bareos-dir.d/pool/MNG-Full.conf
>>>> Pool {
>>>> Name = MNG-Full
>>>> Pool Type = Backup
>>>> Recycle = yes # Bareos can automatically
>>>> recycle Volumes
>>>> AutoPrune = yes # Prune expired volumes
>>>> Volume Retention = 7 days # How long should the Full Backups
>>>> be kept? (#06)
>>>> Maximum Volume Bytes = 50G # Limit Volume size to something
>>>> reasonable
>>>> Maximum Volumes = 1000 # Limit number of Volumes in Pool
>>>> Label Format = "MNG-Full-" # Volumes will be labeled
>>>> "Full-<volume-id>"
>>>> }
>>>>
>>>> *Bareos Storage Daemon*
>>>> */etc/bareos/bareos-sd.d/storage/temng02bossd01-sd.conf*
>>>> Storage {
>>>> Name = temng02bossd01-sd
>>>> Maximum Concurrent Jobs = 30
>>>> }
>>>>
>>>> */etc/bareos/bareos-sd.d/device/FileStorage.conf*
>>>> Device {
>>>> Name = FileStorage
>>>> Media Type = File
>>>> Archive Device = /var/lib/bareos/storage
>>>> LabelMedia = yes; # lets Bareos label unlabeled
>>>> media
>>>> Random Access = yes;
>>>> AutomaticMount = yes; # when device opened, read it
>>>> RemovableMedia = no;
>>>> AlwaysOpen = no;
>>>> Description = "File device. A connecting Director must have the same
>>>> Name and MediaType."
>>>> Maximum Concurrent Jobs = 1
>>>> }
>>>>
>>>> Device {
>>>> Name = FileStorage1
>>>> Media Type = File
>>>> Archive Device = /var/lib/bareos/storage
>>>> LabelMedia = yes; # lets Bareos label unlabeled
>>>> media
>>>> Random Access = yes;
>>>> AutomaticMount = yes; # when device opened, read it
>>>> RemovableMedia = no;
>>>> AlwaysOpen = no;
>>>> Description = "File device. A connecting Director must have the same
>>>> Name and MediaType."
>>>> Maximum Concurrent Jobs = 1
>>>> }
>>>> ###### Repeat until FileStorage29
>>>> ### END
>>>>
>>>> Configuration Summary :
>>>> - I used disk as the backup device and all device pointing to same
>>>> folder which is /var/lib/bareos/storage
>>>> - I already set max concurrent job to all components that need to be
>>>> adjusted, such as Director, Device, Storage, Client.
>>>> - I created an empty pool with maximum volumes 1000, but when I run
>>>> 20++ concurrent job, it only create 19 appendable volumes and got this
>>>> error where it said it cannot find any appendable volume.
>>>> - I'm pretty sure that I already configured concurrent job properly
>>>> because if I force bareos to recycle old volume that already expired,
>>>> concurrent job can be run more than 20 job without any issue.
>>>> - So I can assume that bareos not creating new appendable volume even
>>>> though that i still have a lot of disk space and haven't reach the max
>>>> volumes in the pool.
>>>>
>>>> Question :
>>>> - Is there some configuration that I missing here that make bareos not
>>>> creating new volume?
>>>> - Is this a limitation inside the Bareos?
>>>>
>>>> Kindly inform me if you guys need more information, I will like to
>>>> provide those information in order to solve this blocker.
>>>>
>>>> Thank you.
>>>>
>>>> Regards,
>>>>
>>>> Reqi Rangga
>>>>
>>>>
--
You received this message because you are subscribed to the Google Groups
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/bareos-users/73648f29-2210-4cb8-a273-7c64a815ca28n%40googlegroups.com.