On Thursday, April 21, 2016 at 4:01:58 PM UTC+2, Oliver Hoffmann wrote:
> Hi list,
> 
> 
> something weired is happening. A couple of Jobs start at 1 AM and then the SD 
> crashes. In bactrace I see things like this:
> 
> threadid=0x7f9d62ffd700 JobId=80 JobStatus=C jcr=0x7f9d3c001078 
> name=client1-differential.2016-03-25_01.00.00_16
> threadid=0x7f9d62ffd700 killable=0 JobId=80 JobStatus=C jcr=0x7f9d3c001078 
> name=client1-differential.2016-03-25_01.00.00_16
>         use_count=1
>         JobType=B JobLevel=F
>         sched_time=25-Mar-2016 02:52 start_time=01-Jan-1970 01:00
>         end_time=01-Jan-1970 01:00 wait_time=01-Jan-1970 01:00
>         db=(nil) db_batch=(nil) batch_started=0
> 
> 01-Jan-1970? I checked hwclock and ntp. All good. Where is that date comming 
> from? I saw that people had that bug years and years ago (in bacula).
> 
> I think the SD gets somehow confused due to my unusual configuration. Thing 
> is that I want to have one Volume File for each Job. Thus 20 Devices.
> 
> My sd.conf:
> 
> Storage {
>   Allow Bandwidth Bursting = no
>   Client Connect Wait = 3 minute
>   Collect Device Statistics = yes
>   Collect Job Statistics = yes
>   Description = vampire3 Storage Daemon
>   FD Connect Timeout = 3 minute
>   File Device Concurrent Read = yes
>   Heartbeat Interval = 1 minute
>   Maximum Concurrent Jobs = 20
>   Name = vampire3-sd
>   Statistics Collect Interval = 5 minute
> 
> Director {
>   Name = my-dir
>   Password = "mypass"
> }
> 
> Director {
>   Monitor = yes
>   Name = my-dir
>   Password = "mypass"
> }
> 
> Device {
>   AlwaysOpen = no
>   Archive Device = /mnt/raid/bareos/volumes
>   Automatic Mount = yes
>   Collect Statistics = yes
>   Description = Device-1 RAID
>   Device Type = File
>   Label Media = yes
>   Maximum Concurrent Jobs = 20
>   Maximum Open Volumes = 20
>   Maximum Open Wait = 5 minute
>   Media Type = File
>   Name = V3-RAID-1
>   Random Access = yes
>   Removable Media = no
> }
> 
> .
> .
> .
> (continues till Device-20)
> 
> 
> In the dir.conf I have Storage definitions like this:
> 
> Storage {
>   Address = 10.10.10.11
>   Allow Compression = yes
>   Collect Statistics = yes
>   Description = RAID on 10 network.
>   Device = V3-RAID-1
>   Device = V3-RAID-2
>   Device = V3-RAID-3
>   Device = V3-RAID-4
>   Device = V3-RAID-5
>   Device = V3-RAID-6
>   Device = V3-RAID-7
>   Device = V3-RAID-8
>   Device = V3-RAID-9
>   Device = V3-RAID-10
>   Device = V3-RAID-11
>   Device = V3-RAID-12
>   Device = V3-RAID-13
>   Device = V3-RAID-14
>   Device = V3-RAID-15
>   Device = V3-RAID-16
>   Device = V3-RAID-17
>   Device = V3-RAID-18
>   Device = V3-RAID-19
>   Device = V3-RAID-20
>   Enabled = yes
>   Heartbeat Interval = 1 minute
>   Maximum Concurrent Jobs = 20
>   Maximum Concurrent Read Jobs = 30
>   Media Type = File
>   Name = V3-RAID-10
>   Password = "my-pass"
>   Protocol = Native
>   Port = 9103
> }
> 
> 
> A Pool looks like this:
> 
> Pool {
>   Action On Purge = Truncate
>   Auto Prune = yes
>   Catalog = MyCatalog
>   Catalog Files = yes
>   Description = Pool for Full backups
>   File Retention = 365 days
>   Job Retention = 365 days
>   Label Format = "${Client}-${Year}-${Month}-${Day}-${JobId}"
>   Maximum Volume Jobs = 1
>   Name = V3-Full-RAID
>   Pool Type = Backup
>   Recycle = No
>   Use Catalog = yes
>   Volume Retention = 16 days
> }
> 
> 
> Nothing special about Schedules:
> 
> Schedule {
>   Description = Daily backup schedule
>   Enabled = yes
>   Name = "Daily"
>   Run = sun-fri at 01:00
> }
> 
> 
> After a crash I need to restart the daemons and start the Jobs by hand.
> 
> Any clues?
> 
> Thanks for hints!
> 
> Regards,
> 
> Oliver

Thank you. I'll carry on with a bug report then.

Regards,

Oliver

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to