I don't have a storage class - here's the device definition:

Device { 
 Name = s3
 MediaType = "s3" 
 ArchiveDevice = Object Storage 
 DeviceType = dplcompat 
 Device Options = "iothreads=4" 
                  ",ioslots=2" 
                  ",chunksize=262144000" 
                  ",program=/usr/lib/bareos/scripts/s3cmd-wrapper.sh" 
                  ",s3cfg=/etc/bareos/s3cmd.cfg" 
                  ",bucket=test-backup-pool" 
 LabelMedia = yes 
 RandomAccess = yes 
 AutomaticMount = yes 
 RemovableMedia = no 
 AlwaysOpen = no 
}


On Tuesday, October 28, 2025 at 6:46:47 AM UTC-4 Bruno Friedmann 
(bruno-at-bareos) wrote:

> Hi James,
>
> What you didn't shown here is the configuration of you device
>
> and maybe you have an option line that create an "invisible error" 
> For example with a local s3minio server and the default config example 
> file 
>
> ```
>   Device Options = "iothreads=4"
>                    ",ioslots=2"
>                    ",chunksize=262144000"
>                    ",program=s3cmd-wrapper.sh"
>                    ",s3cfg=/etc/bareos/s3cfg"
>                    ",bucket=bareos"
>                    ",storage_class=STANDARD_IA"
> ```
>
> won't work because minio won't accept the storage_class so I can get 
> whatever test with s3cmd this error will be only visible during the label 
> of the volume with a returned 403 
>
> (this can be visible if you activate debuglevel and trace on the storage 
> with)
> setdebug level=500 trace=1 timestamp=1 storage=dplcompat 
>
> So the fix is easy, remove storage_class from the device configuration or 
> found the right value that need to be there.
>
> With the trace activated you will have detailed informations available and 
> maybe understand  what's going wrong in your setup.
>
> My bet is still that you have one option in bareos-sd device which is not 
> present in the s3cfg which are failing.
>
> On Monday, 27 October 2025 at 17:58:00 UTC+1 James Pulver wrote:
>
>> I just added 
>> # Log the expanded command to /tmp
>>     echo "$s3cmd_prog ${s3cmd_common_args[*]} $*" >> 
>> /tmp/run_s3cmd_command.log
>>
>> to the s3command wrapper script.
>>
>> This shows that the last thing put into that command is
>> /usr/bin/s3cmd --no-progress --config /etc/bareos/s3cmd.cfg info 
>> s3://test-backup-pool
>>
>> I then tested that command doing
>> sudo -u bareos -s
>> /usr/bin/s3cmd --no-progress --config /etc/bareos/s3cmd.cfg info 
>> s3://test-back
>> up-pool
>> Which immediately returns:
>> s3://test-backup-pool/ (bucket): 
>>   Location:  default 
>>   Payer:     BucketOwner 
>>   Ownership: none 
>>   Versioning:none 
>>   Expiration rule: none 
>>   Block Public Access: none 
>>   Policy:    none 
>>   CORS:      none 
>>   ACL:       usr_0001: FULL_CONTROL
>>
>> So as far as I can tell, my /etc/bareos/s3cmd.cfg is correct, and I can 
>> run s3cmd using that config as the bareos user successfully. But the actual 
>> backup just hangs after that command - nothing else happens that I can see. 
>> I'm not sure what it's waiting for to get back? Or what else to check.
>> On Thursday, October 23, 2025 at 4:48:21 AM UTC-4 Bruno Friedmann 
>> (bruno-at-bareos) wrote:
>>
>>> Hi James, as I said previously, seems your /bareoss3test.sh has all the 
>>> needed configuration setup and something was not ported into your bareos 
>>> configuration, otherwise the label should also work.
>>>
>>> On Wednesday, 22 October 2025 at 17:48:56 UTC+2 James Pulver wrote:
>>>
>>>> As far as I can tell, yes I've completed all the steps.
>>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh testconnection
>>>>
>>>> s3://test-backup-pool/ (bucket):
>>>>    Location:  default
>>>>    Payer:     BucketOwner
>>>>    Ownership: none
>>>>    Versioning:none
>>>>    Expiration rule: none
>>>>    Block Public Access: none
>>>>    Policy:    none
>>>>    CORS:      none
>>>>    ACL:       0001: FULL_CONTROL
>>>> [root@lnx bareos]# echo "Hello, World!" | sudo -u bareos 
>>>> ./bareoss3test.sh upload test 0000
>>>> upload: '<stdin>' -> 's3://test-backup-pool/test/0000' (0 bytes in 0.3 
>>>> seconds, -1.00 B/s) [1 of 1]
>>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh list test
>>>> 0000 14
>>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh stat test 0000
>>>> 14
>>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh download test 0000
>>>> Hello, World!
>>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh remove test 0000
>>>> delete: 's3://test-backup-pool/test/0000'
>>>>
>>>> These all succeeded as shown. Manual label fails however from bconsole. 
>>>>
>>>> On Monday, October 20, 2025 at 4:14:01 AM UTC-4 Bruno Friedmann 
>>>> (bruno-at-bareos) wrote:
>>>>
>>>>> Hi James,
>>>>>
>>>>> Did you really fully completed all the steps describe in the 
>>>>> discussion ?
>>>>> Because you shouldn't have such returns afterwards
>>>>>
>>>>> Running ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011" 
>>>>> "0000"" returned 1
>>>>>
>>>>> Please carefully recheck all the steps, and try manually to create 
>>>>> object once this works
>>>>> doing a label manually in Bareos should also works (volume created and 
>>>>> identified).
>>>>>
>>>>> Then the backup should run without any glitches.
>>>>>
>>>>> It's certainly just a question of a details, but which one ;-) 
>>>>>
>>>>>
>>>>> On Thursday, 16 October 2025 at 22:01:29 UTC+2 James Pulver wrote:
>>>>>
>>>>>>
>>>>>> Here's a sanitized version of what I get:
>>>>>> bareos-sd (50): stored/dir_cmd.cc:399-0 level=130 trace=1 timestamp=0 
>>>>>> tracefilename=/var/lib/bareos/bareos-sd.trace
>>>>>> bareos-sd (100): lib/jcr.cc:378-0 Destruct JobControlRecord
>>>>>> bareos-sd (100): lib/jcr.cc:268-0 FreeCommonJcr: 7fbc7c05daa0 
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:89-0 Destruct 
>>>>>> TlsOpenSslPrivate
>>>>>> bareos-sd (100): lib/bsock.cc:137-0 Destruct BareosSocket
>>>>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct 
>>>>>> TlsOpenSslPrivate
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp 
>>>>>> filedescriptor: <5>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: 
>>>>>> <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer: 
>>>>>> <false>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb. 
>>>>>> identitiy: R_DIRECTOR bareos-dir.
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb. 
>>>>>> result: 32.
>>>>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>>>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Director 
>>>>>> bareos-dir calling
>>>>>> bareos-sd (110): stored/socket_server.cc:113-0 Got a DIR connection 
>>>>>> at 16-Oct-2025 15:50:40
>>>>>> bareos-sd (100): lib/jcr.cc:185-0 Construct JobControlRecord
>>>>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5 
>>>>>> <1910067158.1760644240@R_STORAGE::bareos-sd> ssl=2
>>>>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth 
>>>>>> cram-md5 <273648795.1760644240@R_DIRECTOR::bareos-dir> ssl=2
>>>>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> 
>>>>>> - challenge_name: <R_DIRECTOR::bareos-dir>
>>>>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge: 
>>>>>> a/xIaRRst6sme9ROL+/zHC
>>>>>> bareos-sd (90): stored/dir_cmd.cc:276-0 Message channel init 
>>>>>> completed.
>>>>>> bareos-sd (100): stored/job.cc:90-0 <dird: JobId=283 
>>>>>> job=tmp-s3.2025-10-16_15.50.38_35 job_name=tmp-s3 client_name=lnx100-fd 
>>>>>> type=66 level=70 FileSet=tmp NoAttr=0 SpoolAttr=0 
>>>>>> FileSetMD5=1C+cV9JM99s729My3DVDfD SpoolData=0 PreferMountedVols=1 
>>>>>> SpoolSize=0 rerunning=0 VolSessionId=0 VolSessionTime=0 Quota=0 
>>>>>> Protocol=0 
>>>>>> BackupFormat=Native
>>>>>> bareos-sd (100): stored/job.cc:109-0 rerunning=0 VolSesId=0 
>>>>>> VolSesTime=0 Protocol=0
>>>>>> bareos-sd (50): stored/job.cc:155-283 Quota set as 0
>>>>>> bareos-sd (50): stored/job.cc:166-283 >dird jid=283: 3000 OK Job 
>>>>>> SDid=4 SDtime=1760633557 Authorization=
>>>>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct 
>>>>>> TlsOpenSslPrivate
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp 
>>>>>> filedescriptor: <7>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: 
>>>>>> <>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer: 
>>>>>> <false>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb. 
>>>>>> identitiy: R_JOB tmp-s3.2025-10-16_15.50.38_35.
>>>>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb. 
>>>>>> result: 39.
>>>>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>>>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Start Job 
>>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>>> bareos-sd (110): stored/socket_server.cc:101-0 Got a FD connection at 
>>>>>> 16-Oct-2025 15:50:40
>>>>>> bareos-sd (50): stored/fd_cmds.cc:116-0 Found Job 
>>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5 
>>>>>> <178672579.1760644240@R_STORAGE::bareos-sd> ssl=2
>>>>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth 
>>>>>> cram-md5 <710187007.1760644240@R_CLIENT::lnxcmp-fd> ssl=2
>>>>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> 
>>>>>> - challenge_name: <R_CLIENT::lnxcmp-fd>
>>>>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge: 
>>>>>> S0/TKR/RWX/Ju7t327/y0D
>>>>>> bareos-sd (50): stored/job.cc:199-283 tmp-s3.2025-10-16_15.50.38_35 
>>>>>> waiting 1800 sec for FD to contact SD key=
>>>>>> bareos-sd (50): stored/fd_cmds.cc:141-0 OK Authentication jid=283 Job 
>>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>>> bareos-sd (50): stored/job.cc:213-283 Auth=1 canceled=0
>>>>>> bareos-sd (120): stored/fd_cmds.cc:169-283 Start run 
>>>>>> Job=tmp-s3.2025-10-16_15.50.38_35
>>>>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append open session
>>>>>> bareos-sd (120): stored/fd_cmds.cc:291-283 Append open session: 
>>>>>> append open session
>>>>>> bareos-sd (110): stored/fd_cmds.cc:302-283 >filed: 3000 OK open 
>>>>>> ticket = 4
>>>>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append data 4
>>>>>> bareos-sd (120): stored/fd_cmds.cc:255-283 Append data: append data 4
>>>>>> bareos-sd (110): stored/fd_cmds.cc:257-283 <filed: append data 4
>>>>>> bareos-sd (100): lib/bsock.cc:90-283 Copy Contructor BareosSocket
>>>>>> bareos-sd (100): stored/block.cc:137-283 created new block of 
>>>>>> blocksize 1048576 (dev->max_block_size)
>>>>>> bareos-sd (50): stored/askdir.cc:211-283 DirFindNextAppendableVolume: 
>>>>>> reserved=1 Vol=
>>>>>> bareos-sd (50): stored/askdir.cc:261-283 >dird CatReq 
>>>>>> Job=tmp-s3.2025-10-16_15.50.38_35 FindMedia=1 pool_name=Full 
>>>>>> media_type=s3 
>>>>>> unwanted_volumes=
>>>>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK 
>>>>>> VolName=Full-0011 VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 
>>>>>> VolMounts=0 
>>>>>> VolErrors=0 VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0 
>>>>>> VolStatus=Append Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0 
>>>>>> VolReadTime=0 VolWriteTime=0 EndFile=0 EndBlock=0 LabelType=0 MediaId=11 
>>>>>> EncryptionKey= MinBlocksize=0 MaxBlocksize=0
>>>>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true 
>>>>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>>>>> bareos-sd (50): stored/askdir.cc:151-283 setting 
>>>>>> dcr->VolMinBlocksize(0) to vol.VolMinBlocksize(0)
>>>>>> bareos-sd (50): stored/askdir.cc:154-283 setting 
>>>>>> dcr->VolMaxBlocksize(0) to vol.VolMaxBlocksize(0)
>>>>>> bareos-sd (50): stored/askdir.cc:272-283 Call reserve_volume for 
>>>>>> write. Vol=Full-0011
>>>>>> bareos-sd (50): stored/askdir.cc:279-283 DirFindNextAppendableVolume 
>>>>>> return true. vol=Full-0011
>>>>>> bareos-sd (100): stored/append.cc:272-283 Start append data. res=1
>>>>>> bareos-sd (100): stored/acquire.cc:418-283 acquire_append device is 
>>>>>> dplcompat
>>>>>> bareos-sd (100): stored/mount.cc:608-283 No swap_dev set
>>>>>> bareos-sd (50): stored/askdir.cc:186-283 >dird CatReq 
>>>>>> Job=tmp-s3.2025-10-16_15.50.38_35 GetVolInfo VolName=Full-0011 write=1
>>>>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK 
>>>>>> VolName=Full-0011 VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 
>>>>>> VolMounts=0 
>>>>>> VolErrors=0 VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0 
>>>>>> VolStatus=Append Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0 
>>>>>> VolReadTime=0 VolWriteTime=0 EndFile=0 EndBlock=0 LabelType=0 MediaId=11 
>>>>>> EncryptionKey= MinBlocksize=0 MaxBlocksize=0
>>>>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true 
>>>>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>>>>> bareos-sd (50): stored/askdir.cc:151-283 setting 
>>>>>> dcr->VolMinBlocksize(0) to vol.VolMinBlocksize(0)
>>>>>> bareos-sd (50): stored/askdir.cc:154-283 setting 
>>>>>> dcr->VolMaxBlocksize(0) to vol.VolMaxBlocksize(0)
>>>>>> bareos-sd (100): stored/autochanger.cc:125-283 Device "s3" 
>>>>>> (ObjectStorage) is not attached to an autochanger
>>>>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896 
>>>>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011 
>>>>>> mode=OPEN_READ_WRITE
>>>>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice 
>>>>>> mode=OPEN_READ_WRITE
>>>>>> bareos-sd (100): stored/dev.cc:593-283 open archive: 
>>>>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>>>>> bareos-sd (120): backends/dplcompat_device.cc:272-283 
>>>>>> CheckRemoteConnection called
>>>>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection 
>>>>>> called
>>>>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection 
>>>>>> returned 0
>>>>>> == Output ==
>>>>>> s3://test-backup-pool/ (bucket):
>>>>>>    Location:  default
>>>>>>    Payer:     BucketOwner
>>>>>>    Ownership: none
>>>>>>    Versioning:none
>>>>>>    Expiration rule: none
>>>>>>    Block Public Access: none
>>>>>>    Policy:    none
>>>>>>    CORS:      none
>>>>>>    ACL:       s3_0001: FULL_CONTROL
>>>>>> ============
>>>>>> bareos-sd (100): backends/chunked_device.cc:106-283 New allocated 
>>>>>> buffer of 262144000 bytes at 7fbc5c5ff010
>>>>>> bareos-sd (120): backends/dplcompat_device.cc:332-283 Reading chunk 
>>>>>> Full-0011/0000
>>>>>> bareos-sd (130): backends/crud_storage.cc:245-283 stat Full-0011/0000 
>>>>>> called
>>>>>> bareos-sd (130): backends/crud_storage.cc:253-283 stat returned 1
>>>>>> == Output ==
>>>>>> ============
>>>>>> bareos-sd (110): backends/crud_storage.cc:260-283 stat returned 1
>>>>>> bareos-sd (100): backends/dplcompat_device.cc:339-283 Running 
>>>>>> ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011" "0000"" 
>>>>>> returned 1
>>>>>> bareos-sd (100): stored/dev.cc:605-283 open failed: stored/dev.cc:600 
>>>>>> Could not open: ObjectStorage/Full-0011
>>>>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=-1 opened
>>>>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762566 fd=-1
>>>>>> bareos-sd (100): stored/dev.cc:453-283 setting minblocksize to 64512, 
>>>>>> maxblocksize to label_block_size=64512, on device "s3" (ObjectStorage)
>>>>>> bareos-sd (100): stored/block.cc:137-283 created new block of 
>>>>>> blocksize 64512 (dev->max_block_size)
>>>>>> bareos-sd (100): stored/dev.cc:467-283 created new block of buf_len: 
>>>>>> 64512 on device "s3" (ObjectStorage)
>>>>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896 
>>>>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011 
>>>>>> mode=OPEN_READ_WRITE
>>>>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice 
>>>>>> mode=OPEN_READ_WRITE
>>>>>> bareos-sd (100): stored/dev.cc:593-283 open archive: 
>>>>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>>>>> bareos-sd (120): backends/dplcompat_device.cc:272-283 
>>>>>> CheckRemoteConnection called
>>>>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection 
>>>>>> called
>>>>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection 
>>>>>> returned 0
>>>>>> == Output ==
>>>>>> s3://test-backup-pool/ (bucket):
>>>>>>    Location:  default
>>>>>>    Payer:     BucketOwner
>>>>>>    Ownership: none
>>>>>>    Versioning:none
>>>>>>    Expiration rule: none
>>>>>>    Block Public Access: none
>>>>>>    Policy:    none
>>>>>>    CORS:      none
>>>>>>    ACL:       s3_0001: FULL_CONTROL
>>>>>> ============
>>>>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=0 opened
>>>>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762346 fd=0
>>>>>> bareos-sd (130): stored/label.cc:520-283 Start CreateVolumeLabel()
>>>>>> bareos-sd (100): stored/dev.cc:879-283 Clear volhdr vol=
>>>>>>
>>>>>> Volume Label:
>>>>>> Id                : Bareos 2.0 immortal
>>>>>> VerNo             : 20
>>>>>> VolName           : Full-0011
>>>>>> PrevVolName       : 
>>>>>> VolFile           : 0
>>>>>> LabelType         : VOL_LABEL
>>>>>> LabelSize         : 0
>>>>>> PoolName          : Full
>>>>>> MediaType         : s3
>>>>>> PoolType          : Backup
>>>>>> HostName          : bareos
>>>>>> Date label written: 16-Oct-2025 15:50
>>>>>> bareos-sd (130): stored/label.cc:410-283 Wrote label of 194 bytes to 
>>>>>> "s3" (ObjectStorage)
>>>>>> bareos-sd (130): stored/label.cc:415-283 Call WriteBlockToDev()
>>>>>> bareos-sd (130): stored/label.cc:423-283  Wrote block to device
>>>>>> bareos-sd (100): backends/chunked_device.cc:345-283 Enqueueing chunk 
>>>>>> 0 of volume Full-0011 (230 bytes)
>>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>>>>> IO-thread threadid=0x00007fbc7affd640
>>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>>>>> IO-thread threadid=0x00007fbc7b7fe640
>>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>>>>> IO-thread threadid=0x00007fbc7a7fc640
>>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>>>>> IO-thread threadid=0x00007fbc79ffb640
>>>>>> bareos-sd (100): backends/chunked_device.cc:361-283 Allocated chunk 
>>>>>> io request of 48 bytes at 7fbc7c046930
>>>>>> bareos-sd (100): backends/chunked_device.cc:1082-283 storage is 
>>>>>> pending, as there are queued write requests for previous volumes.
>>>>>> bareos-sd (100): backends/chunked_device.cc:417-0 Flushing chunk 0 of 
>>>>>> volume Full-0011 by thread 0x00007fbc7affd640
>>>>>> bareos-sd (120): backends/dplcompat_device.cc:285-0 Flushing chunk 
>>>>>> Full-0011/0000
>>>>>> bareos-sd (100): backends/chunked_device.cc:226-0 Creating inflight 
>>>>>> file /var/lib/bareos/Full-0011@0000%inflight for volume Full-0011, chunk >>>>>> 0
>>>>>> bareos-sd (100): backends/dplcompat_device.cc:290-0 Could not acquire 
>>>>>> inflight lease for Full-0011/0000
>>>>>> bareos-sd (100): backends/chunked_device.cc:449-0 Enqueueing chunk 0 
>>>>>> of volume Full-0011 for retry of upload later
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 
>>>>>> is pending, as there are queued write requests
>>>>>>
>>>>>> Not sure what to make of it except it still seems like it cannot 
>>>>>> actually write out an object to s3. 
>>>>>> On Thursday, October 16, 2025 at 6:19:31 AM UTC-4 Andreas Rogge wrote:
>>>>>>
>>>>>>> Hi James, 
>>>>>>>
>>>>>>> Am 13.10.25 um 17:59 schrieb James Pulver: 
>>>>>>> > I've been trying for weeks to get this to work. I now have managed 
>>>>>>> to 
>>>>>>> > set up the dplcompat plugin, and have been able to follow 
>>>>>>> > troubleshooting like 
>>>>>>> https://github.com/bareos/bareos/discussions/2063 
>>>>>>> > from the command line - CLI commands all work. But when I try and 
>>>>>>> run a 
>>>>>>> > backup to the s3 storage device in BAREOS it just hangs. No errors 
>>>>>>> > noted, and status in director shows the job running forever, and 
>>>>>>> status 
>>>>>>> > on the storage shows 
>>>>>>> > 
>>>>>>> > Device status: 
>>>>>>> > 
>>>>>>> > Device "obj" (ObjectStorage) is being acquired with: 
>>>>>>> >     Volume:      Full-0011 
>>>>>>> >     Pool:        Full 
>>>>>>> >     Media type:  obj 
>>>>>>> > Backend connection is working. 
>>>>>>> This means the backend connection test was successful. 
>>>>>>> > Inflight chunks: 0 
>>>>>>> > Pending IO flush requests: 
>>>>>>> >    /Full-0011/0000 - 230 (try=2) 
>>>>>>> This means chunk 0000 (i.e. the first part of the volume) should be 
>>>>>>> written to the backend storage. Its size is 230 bytes, so probably 
>>>>>>> the 
>>>>>>> volume header. This operation failed for some reason and we are now 
>>>>>>> in 
>>>>>>> try 2. 
>>>>>>>
>>>>>>> > It never progresses. I'm really not sure how to proceed. 
>>>>>>> I think the way to go here is to enable debug logging and tracing in 
>>>>>>> the 
>>>>>>> SD so you'll see what happens with the backend calls. 
>>>>>>>
>>>>>>> You can just do "setdebug trace=1 storage=your-storage level=130" in 
>>>>>>> bconsole. It will print the path to the trace-file that will be 
>>>>>>> created 
>>>>>>> on the SD. 
>>>>>>> In that tracefile you should be able to see what is happening 
>>>>>>> between 
>>>>>>> Bareos and your script and maybe figure out where it fails. 
>>>>>>> You can also send it here (or to me personally, if you don't want to 
>>>>>>> share that publicly) and I'll see if I can help. 
>>>>>>>
>>>>>>> Best Regards, 
>>>>>>> Andreas 
>>>>>>> -- 
>>>>>>> Andreas Rogge [email protected] 
>>>>>>> Bareos GmbH & Co. KG Phone: +49 221-630693-86 <+49%20221%2063069386> 
>>>>>>> http://www.bareos.com 
>>>>>>>
>>>>>>> Sitz der Gesellschaft: Köln | Amtsgericht Köln: HRA 29646 
>>>>>>> Komplementär: Bareos Verwaltungs-GmbH 
>>>>>>> Geschäftsführer: Stephan Dühr, Jörg Steffens, Philipp Storz 
>>>>>>>
>>>>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/d/msgid/bareos-users/926220b1-ab79-465f-899f-432ed5b6c292n%40googlegroups.com.

Reply via email to