Hi James,
What you didn't shown here is the configuration of you device
and maybe you have an option line that create an "invisible error"
For example with a local s3minio server and the default config example file
```
Device Options = "iothreads=4"
",ioslots=2"
",chunksize=262144000"
",program=s3cmd-wrapper.sh"
",s3cfg=/etc/bareos/s3cfg"
",bucket=bareos"
",storage_class=STANDARD_IA"
```
won't work because minio won't accept the storage_class so I can get
whatever test with s3cmd this error will be only visible during the label
of the volume with a returned 403
(this can be visible if you activate debuglevel and trace on the storage
with)
setdebug level=500 trace=1 timestamp=1 storage=dplcompat
So the fix is easy, remove storage_class from the device configuration or
found the right value that need to be there.
With the trace activated you will have detailed informations available and
maybe understand what's going wrong in your setup.
My bet is still that you have one option in bareos-sd device which is not
present in the s3cfg which are failing.
On Monday, 27 October 2025 at 17:58:00 UTC+1 James Pulver wrote:
> I just added
> # Log the expanded command to /tmp
> echo "$s3cmd_prog ${s3cmd_common_args[*]} $*" >>
> /tmp/run_s3cmd_command.log
>
> to the s3command wrapper script.
>
> This shows that the last thing put into that command is
> /usr/bin/s3cmd --no-progress --config /etc/bareos/s3cmd.cfg info
> s3://test-backup-pool
>
> I then tested that command doing
> sudo -u bareos -s
> /usr/bin/s3cmd --no-progress --config /etc/bareos/s3cmd.cfg info
> s3://test-back
> up-pool
> Which immediately returns:
> s3://test-backup-pool/ (bucket):
> Location: default
> Payer: BucketOwner
> Ownership: none
> Versioning:none
> Expiration rule: none
> Block Public Access: none
> Policy: none
> CORS: none
> ACL: usr_0001: FULL_CONTROL
>
> So as far as I can tell, my /etc/bareos/s3cmd.cfg is correct, and I can
> run s3cmd using that config as the bareos user successfully. But the actual
> backup just hangs after that command - nothing else happens that I can see.
> I'm not sure what it's waiting for to get back? Or what else to check.
> On Thursday, October 23, 2025 at 4:48:21 AM UTC-4 Bruno Friedmann
> (bruno-at-bareos) wrote:
>
>> Hi James, as I said previously, seems your /bareoss3test.sh has all the
>> needed configuration setup and something was not ported into your bareos
>> configuration, otherwise the label should also work.
>>
>> On Wednesday, 22 October 2025 at 17:48:56 UTC+2 James Pulver wrote:
>>
>>> As far as I can tell, yes I've completed all the steps.
>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh testconnection
>>>
>>> s3://test-backup-pool/ (bucket):
>>> Location: default
>>> Payer: BucketOwner
>>> Ownership: none
>>> Versioning:none
>>> Expiration rule: none
>>> Block Public Access: none
>>> Policy: none
>>> CORS: none
>>> ACL: 0001: FULL_CONTROL
>>> [root@lnx bareos]# echo "Hello, World!" | sudo -u bareos
>>> ./bareoss3test.sh upload test 0000
>>> upload: '<stdin>' -> 's3://test-backup-pool/test/0000' (0 bytes in 0.3
>>> seconds, -1.00 B/s) [1 of 1]
>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh list test
>>> 0000 14
>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh stat test 0000
>>> 14
>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh download test 0000
>>> Hello, World!
>>> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh remove test 0000
>>> delete: 's3://test-backup-pool/test/0000'
>>>
>>> These all succeeded as shown. Manual label fails however from bconsole.
>>>
>>> On Monday, October 20, 2025 at 4:14:01 AM UTC-4 Bruno Friedmann
>>> (bruno-at-bareos) wrote:
>>>
>>>> Hi James,
>>>>
>>>> Did you really fully completed all the steps describe in the discussion
>>>> ?
>>>> Because you shouldn't have such returns afterwards
>>>>
>>>> Running ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011"
>>>> "0000"" returned 1
>>>>
>>>> Please carefully recheck all the steps, and try manually to create
>>>> object once this works
>>>> doing a label manually in Bareos should also works (volume created and
>>>> identified).
>>>>
>>>> Then the backup should run without any glitches.
>>>>
>>>> It's certainly just a question of a details, but which one ;-)
>>>>
>>>>
>>>> On Thursday, 16 October 2025 at 22:01:29 UTC+2 James Pulver wrote:
>>>>
>>>>>
>>>>> Here's a sanitized version of what I get:
>>>>> bareos-sd (50): stored/dir_cmd.cc:399-0 level=130 trace=1 timestamp=0
>>>>> tracefilename=/var/lib/bareos/bareos-sd.trace
>>>>> bareos-sd (100): lib/jcr.cc:378-0 Destruct JobControlRecord
>>>>> bareos-sd (100): lib/jcr.cc:268-0 FreeCommonJcr: 7fbc7c05daa0
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:89-0 Destruct
>>>>> TlsOpenSslPrivate
>>>>> bareos-sd (100): lib/bsock.cc:137-0 Destruct BareosSocket
>>>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct
>>>>> TlsOpenSslPrivate
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp
>>>>> filedescriptor: <5>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer:
>>>>> <false>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb.
>>>>> identitiy: R_DIRECTOR bareos-dir.
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb.
>>>>> result: 32.
>>>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Director
>>>>> bareos-dir calling
>>>>> bareos-sd (110): stored/socket_server.cc:113-0 Got a DIR connection at
>>>>> 16-Oct-2025 15:50:40
>>>>> bareos-sd (100): lib/jcr.cc:185-0 Construct JobControlRecord
>>>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5
>>>>> <1910067158.1760644240@R_STORAGE::bareos-sd> ssl=2
>>>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth
>>>>> cram-md5 <273648795.1760644240@R_DIRECTOR::bareos-dir> ssl=2
>>>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> -
>>>>> challenge_name: <R_DIRECTOR::bareos-dir>
>>>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge:
>>>>> a/xIaRRst6sme9ROL+/zHC
>>>>> bareos-sd (90): stored/dir_cmd.cc:276-0 Message channel init completed.
>>>>> bareos-sd (100): stored/job.cc:90-0 <dird: JobId=283
>>>>> job=tmp-s3.2025-10-16_15.50.38_35 job_name=tmp-s3 client_name=lnx100-fd
>>>>> type=66 level=70 FileSet=tmp NoAttr=0 SpoolAttr=0
>>>>> FileSetMD5=1C+cV9JM99s729My3DVDfD SpoolData=0 PreferMountedVols=1
>>>>> SpoolSize=0 rerunning=0 VolSessionId=0 VolSessionTime=0 Quota=0
>>>>> Protocol=0
>>>>> BackupFormat=Native
>>>>> bareos-sd (100): stored/job.cc:109-0 rerunning=0 VolSesId=0
>>>>> VolSesTime=0 Protocol=0
>>>>> bareos-sd (50): stored/job.cc:155-283 Quota set as 0
>>>>> bareos-sd (50): stored/job.cc:166-283 >dird jid=283: 3000 OK Job
>>>>> SDid=4 SDtime=1760633557 Authorization=
>>>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct
>>>>> TlsOpenSslPrivate
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp
>>>>> filedescriptor: <7>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: <>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer:
>>>>> <false>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb.
>>>>> identitiy: R_JOB tmp-s3.2025-10-16_15.50.38_35.
>>>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb.
>>>>> result: 39.
>>>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Start Job
>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>> bareos-sd (110): stored/socket_server.cc:101-0 Got a FD connection at
>>>>> 16-Oct-2025 15:50:40
>>>>> bareos-sd (50): stored/fd_cmds.cc:116-0 Found Job
>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5
>>>>> <178672579.1760644240@R_STORAGE::bareos-sd> ssl=2
>>>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth
>>>>> cram-md5 <710187007.1760644240@R_CLIENT::lnxcmp-fd> ssl=2
>>>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> -
>>>>> challenge_name: <R_CLIENT::lnxcmp-fd>
>>>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge:
>>>>> S0/TKR/RWX/Ju7t327/y0D
>>>>> bareos-sd (50): stored/job.cc:199-283 tmp-s3.2025-10-16_15.50.38_35
>>>>> waiting 1800 sec for FD to contact SD key=
>>>>> bareos-sd (50): stored/fd_cmds.cc:141-0 OK Authentication jid=283 Job
>>>>> tmp-s3.2025-10-16_15.50.38_35
>>>>> bareos-sd (50): stored/job.cc:213-283 Auth=1 canceled=0
>>>>> bareos-sd (120): stored/fd_cmds.cc:169-283 Start run
>>>>> Job=tmp-s3.2025-10-16_15.50.38_35
>>>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append open session
>>>>> bareos-sd (120): stored/fd_cmds.cc:291-283 Append open session: append
>>>>> open session
>>>>> bareos-sd (110): stored/fd_cmds.cc:302-283 >filed: 3000 OK open ticket
>>>>> = 4
>>>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append data 4
>>>>> bareos-sd (120): stored/fd_cmds.cc:255-283 Append data: append data 4
>>>>> bareos-sd (110): stored/fd_cmds.cc:257-283 <filed: append data 4
>>>>> bareos-sd (100): lib/bsock.cc:90-283 Copy Contructor BareosSocket
>>>>> bareos-sd (100): stored/block.cc:137-283 created new block of
>>>>> blocksize 1048576 (dev->max_block_size)
>>>>> bareos-sd (50): stored/askdir.cc:211-283 DirFindNextAppendableVolume:
>>>>> reserved=1 Vol=
>>>>> bareos-sd (50): stored/askdir.cc:261-283 >dird CatReq
>>>>> Job=tmp-s3.2025-10-16_15.50.38_35 FindMedia=1 pool_name=Full
>>>>> media_type=s3
>>>>> unwanted_volumes=
>>>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK
>>>>> VolName=Full-0011 VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 VolMounts=0
>>>>> VolErrors=0 VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0
>>>>> VolStatus=Append Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0
>>>>> VolReadTime=0 VolWriteTime=0 EndFile=0 EndBlock=0 LabelType=0 MediaId=11
>>>>> EncryptionKey= MinBlocksize=0 MaxBlocksize=0
>>>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true
>>>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>>>> bareos-sd (50): stored/askdir.cc:151-283 setting
>>>>> dcr->VolMinBlocksize(0) to vol.VolMinBlocksize(0)
>>>>> bareos-sd (50): stored/askdir.cc:154-283 setting
>>>>> dcr->VolMaxBlocksize(0) to vol.VolMaxBlocksize(0)
>>>>> bareos-sd (50): stored/askdir.cc:272-283 Call reserve_volume for
>>>>> write. Vol=Full-0011
>>>>> bareos-sd (50): stored/askdir.cc:279-283 DirFindNextAppendableVolume
>>>>> return true. vol=Full-0011
>>>>> bareos-sd (100): stored/append.cc:272-283 Start append data. res=1
>>>>> bareos-sd (100): stored/acquire.cc:418-283 acquire_append device is
>>>>> dplcompat
>>>>> bareos-sd (100): stored/mount.cc:608-283 No swap_dev set
>>>>> bareos-sd (50): stored/askdir.cc:186-283 >dird CatReq
>>>>> Job=tmp-s3.2025-10-16_15.50.38_35 GetVolInfo VolName=Full-0011 write=1
>>>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK
>>>>> VolName=Full-0011 VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 VolMounts=0
>>>>> VolErrors=0 VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0
>>>>> VolStatus=Append Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0
>>>>> VolReadTime=0 VolWriteTime=0 EndFile=0 EndBlock=0 LabelType=0 MediaId=11
>>>>> EncryptionKey= MinBlocksize=0 MaxBlocksize=0
>>>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true
>>>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>>>> bareos-sd (50): stored/askdir.cc:151-283 setting
>>>>> dcr->VolMinBlocksize(0) to vol.VolMinBlocksize(0)
>>>>> bareos-sd (50): stored/askdir.cc:154-283 setting
>>>>> dcr->VolMaxBlocksize(0) to vol.VolMaxBlocksize(0)
>>>>> bareos-sd (100): stored/autochanger.cc:125-283 Device "s3"
>>>>> (ObjectStorage) is not attached to an autochanger
>>>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896
>>>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011
>>>>> mode=OPEN_READ_WRITE
>>>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice
>>>>> mode=OPEN_READ_WRITE
>>>>> bareos-sd (100): stored/dev.cc:593-283 open archive:
>>>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>>>> bareos-sd (120): backends/dplcompat_device.cc:272-283
>>>>> CheckRemoteConnection called
>>>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection
>>>>> called
>>>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection
>>>>> returned 0
>>>>> == Output ==
>>>>> s3://test-backup-pool/ (bucket):
>>>>> Location: default
>>>>> Payer: BucketOwner
>>>>> Ownership: none
>>>>> Versioning:none
>>>>> Expiration rule: none
>>>>> Block Public Access: none
>>>>> Policy: none
>>>>> CORS: none
>>>>> ACL: s3_0001: FULL_CONTROL
>>>>> ============
>>>>> bareos-sd (100): backends/chunked_device.cc:106-283 New allocated
>>>>> buffer of 262144000 bytes at 7fbc5c5ff010
>>>>> bareos-sd (120): backends/dplcompat_device.cc:332-283 Reading chunk
>>>>> Full-0011/0000
>>>>> bareos-sd (130): backends/crud_storage.cc:245-283 stat Full-0011/0000
>>>>> called
>>>>> bareos-sd (130): backends/crud_storage.cc:253-283 stat returned 1
>>>>> == Output ==
>>>>> ============
>>>>> bareos-sd (110): backends/crud_storage.cc:260-283 stat returned 1
>>>>> bareos-sd (100): backends/dplcompat_device.cc:339-283 Running
>>>>> ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011" "0000""
>>>>> returned 1
>>>>> bareos-sd (100): stored/dev.cc:605-283 open failed: stored/dev.cc:600
>>>>> Could not open: ObjectStorage/Full-0011
>>>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=-1 opened
>>>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762566 fd=-1
>>>>> bareos-sd (100): stored/dev.cc:453-283 setting minblocksize to 64512,
>>>>> maxblocksize to label_block_size=64512, on device "s3" (ObjectStorage)
>>>>> bareos-sd (100): stored/block.cc:137-283 created new block of
>>>>> blocksize 64512 (dev->max_block_size)
>>>>> bareos-sd (100): stored/dev.cc:467-283 created new block of buf_len:
>>>>> 64512 on device "s3" (ObjectStorage)
>>>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896
>>>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011
>>>>> mode=OPEN_READ_WRITE
>>>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice
>>>>> mode=OPEN_READ_WRITE
>>>>> bareos-sd (100): stored/dev.cc:593-283 open archive:
>>>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>>>> bareos-sd (120): backends/dplcompat_device.cc:272-283
>>>>> CheckRemoteConnection called
>>>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection
>>>>> called
>>>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection
>>>>> returned 0
>>>>> == Output ==
>>>>> s3://test-backup-pool/ (bucket):
>>>>> Location: default
>>>>> Payer: BucketOwner
>>>>> Ownership: none
>>>>> Versioning:none
>>>>> Expiration rule: none
>>>>> Block Public Access: none
>>>>> Policy: none
>>>>> CORS: none
>>>>> ACL: s3_0001: FULL_CONTROL
>>>>> ============
>>>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=0 opened
>>>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762346 fd=0
>>>>> bareos-sd (130): stored/label.cc:520-283 Start CreateVolumeLabel()
>>>>> bareos-sd (100): stored/dev.cc:879-283 Clear volhdr vol=
>>>>>
>>>>> Volume Label:
>>>>> Id : Bareos 2.0 immortal
>>>>> VerNo : 20
>>>>> VolName : Full-0011
>>>>> PrevVolName :
>>>>> VolFile : 0
>>>>> LabelType : VOL_LABEL
>>>>> LabelSize : 0
>>>>> PoolName : Full
>>>>> MediaType : s3
>>>>> PoolType : Backup
>>>>> HostName : bareos
>>>>> Date label written: 16-Oct-2025 15:50
>>>>> bareos-sd (130): stored/label.cc:410-283 Wrote label of 194 bytes to
>>>>> "s3" (ObjectStorage)
>>>>> bareos-sd (130): stored/label.cc:415-283 Call WriteBlockToDev()
>>>>> bareos-sd (130): stored/label.cc:423-283 Wrote block to device
>>>>> bareos-sd (100): backends/chunked_device.cc:345-283 Enqueueing chunk 0
>>>>> of volume Full-0011 (230 bytes)
>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new
>>>>> IO-thread threadid=0x00007fbc7affd640
>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new
>>>>> IO-thread threadid=0x00007fbc7b7fe640
>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new
>>>>> IO-thread threadid=0x00007fbc7a7fc640
>>>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new
>>>>> IO-thread threadid=0x00007fbc79ffb640
>>>>> bareos-sd (100): backends/chunked_device.cc:361-283 Allocated chunk io
>>>>> request of 48 bytes at 7fbc7c046930
>>>>> bareos-sd (100): backends/chunked_device.cc:1082-283 storage is
>>>>> pending, as there are queued write requests for previous volumes.
>>>>> bareos-sd (100): backends/chunked_device.cc:417-0 Flushing chunk 0 of
>>>>> volume Full-0011 by thread 0x00007fbc7affd640
>>>>> bareos-sd (120): backends/dplcompat_device.cc:285-0 Flushing chunk
>>>>> Full-0011/0000
>>>>> bareos-sd (100): backends/chunked_device.cc:226-0 Creating inflight
>>>>> file /var/lib/bareos/Full-0011@0000%inflight for volume Full-0011, chunk 0
>>>>> bareos-sd (100): backends/dplcompat_device.cc:290-0 Could not acquire
>>>>> inflight lease for Full-0011/0000
>>>>> bareos-sd (100): backends/chunked_device.cc:449-0 Enqueueing chunk 0
>>>>> of volume Full-0011 for retry of upload later
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011
>>>>> is pending, as there are queued write requests
>>>>>
>>>>> Not sure what to make of it except it still seems like it cannot
>>>>> actually write out an object to s3.
>>>>> On Thursday, October 16, 2025 at 6:19:31 AM UTC-4 Andreas Rogge wrote:
>>>>>
>>>>>> Hi James,
>>>>>>
>>>>>> Am 13.10.25 um 17:59 schrieb James Pulver:
>>>>>> > I've been trying for weeks to get this to work. I now have managed
>>>>>> to
>>>>>> > set up the dplcompat plugin, and have been able to follow
>>>>>> > troubleshooting like
>>>>>> https://github.com/bareos/bareos/discussions/2063
>>>>>> > from the command line - CLI commands all work. But when I try and
>>>>>> run a
>>>>>> > backup to the s3 storage device in BAREOS it just hangs. No errors
>>>>>> > noted, and status in director shows the job running forever, and
>>>>>> status
>>>>>> > on the storage shows
>>>>>> >
>>>>>> > Device status:
>>>>>> >
>>>>>> > Device "obj" (ObjectStorage) is being acquired with:
>>>>>> > Volume: Full-0011
>>>>>> > Pool: Full
>>>>>> > Media type: obj
>>>>>> > Backend connection is working.
>>>>>> This means the backend connection test was successful.
>>>>>> > Inflight chunks: 0
>>>>>> > Pending IO flush requests:
>>>>>> > /Full-0011/0000 - 230 (try=2)
>>>>>> This means chunk 0000 (i.e. the first part of the volume) should be
>>>>>> written to the backend storage. Its size is 230 bytes, so probably
>>>>>> the
>>>>>> volume header. This operation failed for some reason and we are now
>>>>>> in
>>>>>> try 2.
>>>>>>
>>>>>> > It never progresses. I'm really not sure how to proceed.
>>>>>> I think the way to go here is to enable debug logging and tracing in
>>>>>> the
>>>>>> SD so you'll see what happens with the backend calls.
>>>>>>
>>>>>> You can just do "setdebug trace=1 storage=your-storage level=130" in
>>>>>> bconsole. It will print the path to the trace-file that will be
>>>>>> created
>>>>>> on the SD.
>>>>>> In that tracefile you should be able to see what is happening between
>>>>>> Bareos and your script and maybe figure out where it fails.
>>>>>> You can also send it here (or to me personally, if you don't want to
>>>>>> share that publicly) and I'll see if I can help.
>>>>>>
>>>>>> Best Regards,
>>>>>> Andreas
>>>>>> --
>>>>>> Andreas Rogge [email protected]
>>>>>> Bareos GmbH & Co. KG Phone: +49 221-630693-86 <+49%20221%2063069386>
>>>>>> http://www.bareos.com
>>>>>>
>>>>>> Sitz der Gesellschaft: Köln | Amtsgericht Köln: HRA 29646
>>>>>> Komplementär: Bareos Verwaltungs-GmbH
>>>>>> Geschäftsführer: Stephan Dühr, Jörg Steffens, Philipp Storz
>>>>>>
>>>>>>
--
You received this message because you are subscribed to the Google Groups
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion visit
https://groups.google.com/d/msgid/bareos-users/21f5def3-3636-49bd-8c0d-9ff5bfdd87d8n%40googlegroups.com.