Hi James, as I said previously, seems your /bareoss3test.sh has all the 
needed configuration setup and something was not ported into your bareos 
configuration, otherwise the label should also work.

On Wednesday, 22 October 2025 at 17:48:56 UTC+2 James Pulver wrote:

> As far as I can tell, yes I've completed all the steps.
> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh testconnection
>
> s3://test-backup-pool/ (bucket):
>    Location:  default
>    Payer:     BucketOwner
>    Ownership: none
>    Versioning:none
>    Expiration rule: none
>    Block Public Access: none
>    Policy:    none
>    CORS:      none
>    ACL:       0001: FULL_CONTROL
> [root@lnx bareos]# echo "Hello, World!" | sudo -u bareos ./bareoss3test.sh 
> upload test 0000
> upload: '<stdin>' -> 's3://test-backup-pool/test/0000' (0 bytes in 0.3 
> seconds, -1.00 B/s) [1 of 1]
> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh list test
> 0000 14
> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh stat test 0000
> 14
> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh download test 0000
> Hello, World!
> [root@lnx bareos]# sudo -u bareos ./bareoss3test.sh remove test 0000
> delete: 's3://test-backup-pool/test/0000'
>
> These all succeeded as shown. Manual label fails however from bconsole. 
>
> On Monday, October 20, 2025 at 4:14:01 AM UTC-4 Bruno Friedmann 
> (bruno-at-bareos) wrote:
>
>> Hi James,
>>
>> Did you really fully completed all the steps describe in the discussion ?
>> Because you shouldn't have such returns afterwards
>>
>> Running ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011" 
>> "0000"" returned 1
>>
>> Please carefully recheck all the steps, and try manually to create object 
>> once this works
>> doing a label manually in Bareos should also works (volume created and 
>> identified).
>>
>> Then the backup should run without any glitches.
>>
>> It's certainly just a question of a details, but which one ;-) 
>>
>>
>> On Thursday, 16 October 2025 at 22:01:29 UTC+2 James Pulver wrote:
>>
>>>
>>> Here's a sanitized version of what I get:
>>> bareos-sd (50): stored/dir_cmd.cc:399-0 level=130 trace=1 timestamp=0 
>>> tracefilename=/var/lib/bareos/bareos-sd.trace
>>> bareos-sd (100): lib/jcr.cc:378-0 Destruct JobControlRecord
>>> bareos-sd (100): lib/jcr.cc:268-0 FreeCommonJcr: 7fbc7c05daa0 
>>> bareos-sd (100): lib/tls_openssl_private.cc:89-0 Destruct 
>>> TlsOpenSslPrivate
>>> bareos-sd (100): lib/bsock.cc:137-0 Destruct BareosSocket
>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct 
>>> TlsOpenSslPrivate
>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp 
>>> filedescriptor: <5>
>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer: 
>>> <false>
>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb. 
>>> identitiy: R_DIRECTOR bareos-dir.
>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb. result: 
>>> 32.
>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Director 
>>> bareos-dir calling
>>> bareos-sd (110): stored/socket_server.cc:113-0 Got a DIR connection at 
>>> 16-Oct-2025 15:50:40
>>> bareos-sd (100): lib/jcr.cc:185-0 Construct JobControlRecord
>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5 
>>> <1910067158.1760644240@R_STORAGE::bareos-sd> ssl=2
>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth cram-md5 
>>> <273648795.1760644240@R_DIRECTOR::bareos-dir> ssl=2
>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> - 
>>> challenge_name: <R_DIRECTOR::bareos-dir>
>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge: 
>>> a/xIaRRst6sme9ROL+/zHC
>>> bareos-sd (90): stored/dir_cmd.cc:276-0 Message channel init completed.
>>> bareos-sd (100): stored/job.cc:90-0 <dird: JobId=283 
>>> job=tmp-s3.2025-10-16_15.50.38_35 job_name=tmp-s3 client_name=lnx100-fd 
>>> type=66 level=70 FileSet=tmp NoAttr=0 SpoolAttr=0 
>>> FileSetMD5=1C+cV9JM99s729My3DVDfD SpoolData=0 PreferMountedVols=1 
>>> SpoolSize=0 rerunning=0 VolSessionId=0 VolSessionTime=0 Quota=0 Protocol=0 
>>> BackupFormat=Native
>>> bareos-sd (100): stored/job.cc:109-0 rerunning=0 VolSesId=0 VolSesTime=0 
>>> Protocol=0
>>> bareos-sd (50): stored/job.cc:155-283 Quota set as 0
>>> bareos-sd (50): stored/job.cc:166-283 >dird jid=283: 3000 OK Job SDid=4 
>>> SDtime=1760633557 Authorization=
>>> bareos-sd (100): lib/bsock.cc:85-0 Construct BareosSocket
>>> bareos-sd (100): lib/tls_openssl_private.cc:58-0 Construct 
>>> TlsOpenSslPrivate
>>> bareos-sd (100): lib/tls_openssl_private.cc:637-0 Set tcp 
>>> filedescriptor: <7>
>>> bareos-sd (100): lib/tls_openssl_private.cc:655-0 Set protocol: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:577-0 Set ca_certfile: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:583-0 Set ca_certdir: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:589-0 Set crlfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:595-0 Set certfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:601-0 Set keyfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:619-0 Set dhfile_: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:643-0 Set cipherlist: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:649-0 Set ciphersuites: <>
>>> bareos-sd (100): lib/tls_openssl_private.cc:625-0 Set Verify Peer: 
>>> <false>
>>> bareos-sd (100): lib/tls_openssl_private.cc:631-0 Set ktls: <false>
>>> bareos-sd (100): lib/tls_openssl_private.cc:501-0 psk_server_cb. 
>>> identitiy: R_JOB tmp-s3.2025-10-16_15.50.38_35.
>>> bareos-sd (100): lib/tls_openssl_private.cc:522-0 psk_server_cb. result: 
>>> 39.
>>> bareos-sd (50): lib/bnet.cc:143-0 TLS server negotiation established.
>>> bareos-sd (110): stored/socket_server.cc:97-0 Conn: Hello Start Job 
>>> tmp-s3.2025-10-16_15.50.38_35
>>> bareos-sd (110): stored/socket_server.cc:101-0 Got a FD connection at 
>>> 16-Oct-2025 15:50:40
>>> bareos-sd (50): stored/fd_cmds.cc:116-0 Found Job 
>>> tmp-s3.2025-10-16_15.50.38_35
>>> bareos-sd (50): lib/cram_md5.cc:106-0 send: auth cram-md5 
>>> <178672579.1760644240@R_STORAGE::bareos-sd> ssl=2
>>> bareos-sd (100): lib/cram_md5.cc:167-0 cram-get received: auth cram-md5 
>>> <710187007.1760644240@R_CLIENT::lnxcmp-fd> ssl=2
>>> bareos-sd (50): lib/cram_md5.cc:61-0 my_name: <R_STORAGE::bareos-sd> - 
>>> challenge_name: <R_CLIENT::lnxcmp-fd>
>>> bareos-sd (99): lib/cram_md5.cc:232-0 sending resp to challenge: 
>>> S0/TKR/RWX/Ju7t327/y0D
>>> bareos-sd (50): stored/job.cc:199-283 tmp-s3.2025-10-16_15.50.38_35 
>>> waiting 1800 sec for FD to contact SD key=
>>> bareos-sd (50): stored/fd_cmds.cc:141-0 OK Authentication jid=283 Job 
>>> tmp-s3.2025-10-16_15.50.38_35
>>> bareos-sd (50): stored/job.cc:213-283 Auth=1 canceled=0
>>> bareos-sd (120): stored/fd_cmds.cc:169-283 Start run 
>>> Job=tmp-s3.2025-10-16_15.50.38_35
>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append open session
>>> bareos-sd (120): stored/fd_cmds.cc:291-283 Append open session: append 
>>> open session
>>> bareos-sd (110): stored/fd_cmds.cc:302-283 >filed: 3000 OK open ticket = 
>>> 4
>>> bareos-sd (110): stored/fd_cmds.cc:210-283 <filed: append data 4
>>> bareos-sd (120): stored/fd_cmds.cc:255-283 Append data: append data 4
>>> bareos-sd (110): stored/fd_cmds.cc:257-283 <filed: append data 4
>>> bareos-sd (100): lib/bsock.cc:90-283 Copy Contructor BareosSocket
>>> bareos-sd (100): stored/block.cc:137-283 created new block of blocksize 
>>> 1048576 (dev->max_block_size)
>>> bareos-sd (50): stored/askdir.cc:211-283 DirFindNextAppendableVolume: 
>>> reserved=1 Vol=
>>> bareos-sd (50): stored/askdir.cc:261-283 >dird CatReq 
>>> Job=tmp-s3.2025-10-16_15.50.38_35 FindMedia=1 pool_name=Full media_type=s3 
>>> unwanted_volumes=
>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK VolName=Full-0011 
>>> VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 VolMounts=0 VolErrors=0 
>>> VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0 VolStatus=Append 
>>> Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0 VolReadTime=0 VolWriteTime=0 
>>> EndFile=0 EndBlock=0 LabelType=0 MediaId=11 EncryptionKey= MinBlocksize=0 
>>> MaxBlocksize=0
>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true 
>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>> bareos-sd (50): stored/askdir.cc:151-283 setting dcr->VolMinBlocksize(0) 
>>> to vol.VolMinBlocksize(0)
>>> bareos-sd (50): stored/askdir.cc:154-283 setting dcr->VolMaxBlocksize(0) 
>>> to vol.VolMaxBlocksize(0)
>>> bareos-sd (50): stored/askdir.cc:272-283 Call reserve_volume for write. 
>>> Vol=Full-0011
>>> bareos-sd (50): stored/askdir.cc:279-283 DirFindNextAppendableVolume 
>>> return true. vol=Full-0011
>>> bareos-sd (100): stored/append.cc:272-283 Start append data. res=1
>>> bareos-sd (100): stored/acquire.cc:418-283 acquire_append device is 
>>> dplcompat
>>> bareos-sd (100): stored/mount.cc:608-283 No swap_dev set
>>> bareos-sd (50): stored/askdir.cc:186-283 >dird CatReq 
>>> Job=tmp-s3.2025-10-16_15.50.38_35 GetVolInfo VolName=Full-0011 write=1
>>> bareos-sd (50): stored/askdir.cc:117-283 <dird 1000 OK VolName=Full-0011 
>>> VolJobs=0 VolFiles=0 VolBlocks=0 VolBytes=0 VolMounts=0 VolErrors=0 
>>> VolWrites=0 MaxVolBytes=53687091200 VolCapacityBytes=0 VolStatus=Append 
>>> Slot=0 MaxVolJobs=0 MaxVolFiles=0 InChanger=0 VolReadTime=0 VolWriteTime=0 
>>> EndFile=0 EndBlock=0 LabelType=0 MediaId=11 EncryptionKey= MinBlocksize=0 
>>> MaxBlocksize=0
>>> bareos-sd (50): stored/askdir.cc:147-283 DoGetVolumeInfo return true 
>>> slot=0 Volume=Full-0011, VolminBlocksize=0 VolMaxBlocksize=0
>>> bareos-sd (50): stored/askdir.cc:151-283 setting dcr->VolMinBlocksize(0) 
>>> to vol.VolMinBlocksize(0)
>>> bareos-sd (50): stored/askdir.cc:154-283 setting dcr->VolMaxBlocksize(0) 
>>> to vol.VolMaxBlocksize(0)
>>> bareos-sd (100): stored/autochanger.cc:125-283 Device "s3" 
>>> (ObjectStorage) is not attached to an autochanger
>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896 
>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011 
>>> mode=OPEN_READ_WRITE
>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice 
>>> mode=OPEN_READ_WRITE
>>> bareos-sd (100): stored/dev.cc:593-283 open archive: 
>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>> bareos-sd (120): backends/dplcompat_device.cc:272-283 
>>> CheckRemoteConnection called
>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection called
>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection 
>>> returned 0
>>> == Output ==
>>> s3://test-backup-pool/ (bucket):
>>>    Location:  default
>>>    Payer:     BucketOwner
>>>    Ownership: none
>>>    Versioning:none
>>>    Expiration rule: none
>>>    Block Public Access: none
>>>    Policy:    none
>>>    CORS:      none
>>>    ACL:       s3_0001: FULL_CONTROL
>>> ============
>>> bareos-sd (100): backends/chunked_device.cc:106-283 New allocated buffer 
>>> of 262144000 bytes at 7fbc5c5ff010
>>> bareos-sd (120): backends/dplcompat_device.cc:332-283 Reading chunk 
>>> Full-0011/0000
>>> bareos-sd (130): backends/crud_storage.cc:245-283 stat Full-0011/0000 
>>> called
>>> bareos-sd (130): backends/crud_storage.cc:253-283 stat returned 1
>>> == Output ==
>>> ============
>>> bareos-sd (110): backends/crud_storage.cc:260-283 stat returned 1
>>> bareos-sd (100): backends/dplcompat_device.cc:339-283 Running 
>>> ""/usr/lib/bareos/scripts/s3cmd-wrapper.sh" stat "Full-0011" "0000"" 
>>> returned 1
>>> bareos-sd (100): stored/dev.cc:605-283 open failed: stored/dev.cc:600 
>>> Could not open: ObjectStorage/Full-0011
>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=-1 opened
>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762566 fd=-1
>>> bareos-sd (100): stored/dev.cc:453-283 setting minblocksize to 64512, 
>>> maxblocksize to label_block_size=64512, on device "s3" (ObjectStorage)
>>> bareos-sd (100): stored/block.cc:137-283 created new block of blocksize 
>>> 64512 (dev->max_block_size)
>>> bareos-sd (100): stored/dev.cc:467-283 created new block of buf_len: 
>>> 64512 on device "s3" (ObjectStorage)
>>> bareos-sd (100): stored/dev.cc:510-283 open dev: type=2080388896 
>>> archive_device_string="s3" (ObjectStorage) vol=Full-0011 
>>> mode=OPEN_READ_WRITE
>>> bareos-sd (100): stored/dev.cc:528-283 call OpenDevice 
>>> mode=OPEN_READ_WRITE
>>> bareos-sd (100): stored/dev.cc:593-283 open archive: 
>>> mode=OPEN_READ_WRITE open(ObjectStorage/Full-0011, 00000002, 0640)
>>> bareos-sd (120): backends/dplcompat_device.cc:272-283 
>>> CheckRemoteConnection called
>>> bareos-sd (130): backends/crud_storage.cc:221-283 test_connection called
>>> bareos-sd (130): backends/crud_storage.cc:229-283 testconnection 
>>> returned 0
>>> == Output ==
>>> s3://test-backup-pool/ (bucket):
>>>    Location:  default
>>>    Payer:     BucketOwner
>>>    Ownership: none
>>>    Versioning:none
>>>    Expiration rule: none
>>>    Block Public Access: none
>>>    Policy:    none
>>>    CORS:      none
>>>    ACL:       s3_0001: FULL_CONTROL
>>> ============
>>> bareos-sd (100): stored/dev.cc:614-283 open dev: disk fd=0 opened
>>> bareos-sd (100): stored/dev.cc:534-283 preserve=20127762346 fd=0
>>> bareos-sd (130): stored/label.cc:520-283 Start CreateVolumeLabel()
>>> bareos-sd (100): stored/dev.cc:879-283 Clear volhdr vol=
>>>
>>> Volume Label:
>>> Id                : Bareos 2.0 immortal
>>> VerNo             : 20
>>> VolName           : Full-0011
>>> PrevVolName       : 
>>> VolFile           : 0
>>> LabelType         : VOL_LABEL
>>> LabelSize         : 0
>>> PoolName          : Full
>>> MediaType         : s3
>>> PoolType          : Backup
>>> HostName          : bareos
>>> Date label written: 16-Oct-2025 15:50
>>> bareos-sd (130): stored/label.cc:410-283 Wrote label of 194 bytes to 
>>> "s3" (ObjectStorage)
>>> bareos-sd (130): stored/label.cc:415-283 Call WriteBlockToDev()
>>> bareos-sd (130): stored/label.cc:423-283  Wrote block to device
>>> bareos-sd (100): backends/chunked_device.cc:345-283 Enqueueing chunk 0 
>>> of volume Full-0011 (230 bytes)
>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>> IO-thread threadid=0x00007fbc7affd640
>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>> IO-thread threadid=0x00007fbc7b7fe640
>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>> IO-thread threadid=0x00007fbc7a7fc640
>>> bareos-sd (100): backends/chunked_device.cc:164-283 Started new 
>>> IO-thread threadid=0x00007fbc79ffb640
>>> bareos-sd (100): backends/chunked_device.cc:361-283 Allocated chunk io 
>>> request of 48 bytes at 7fbc7c046930
>>> bareos-sd (100): backends/chunked_device.cc:1082-283 storage is pending, 
>>> as there are queued write requests for previous volumes.
>>> bareos-sd (100): backends/chunked_device.cc:417-0 Flushing chunk 0 of 
>>> volume Full-0011 by thread 0x00007fbc7affd640
>>> bareos-sd (120): backends/dplcompat_device.cc:285-0 Flushing chunk 
>>> Full-0011/0000
>>> bareos-sd (100): backends/chunked_device.cc:226-0 Creating inflight file 
>>> /var/lib/bareos/Full-0011@0000%inflight for volume Full-0011, chunk 0
>>> bareos-sd (100): backends/dplcompat_device.cc:290-0 Could not acquire 
>>> inflight lease for Full-0011/0000
>>> bareos-sd (100): backends/chunked_device.cc:449-0 Enqueueing chunk 0 of 
>>> volume Full-0011 for retry of upload later
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>> bareos-sd (100): backends/chunked_device.cc:1078-283 volume Full-0011 is 
>>> pending, as there are queued write requests
>>>
>>> Not sure what to make of it except it still seems like it cannot 
>>> actually write out an object to s3. 
>>> On Thursday, October 16, 2025 at 6:19:31 AM UTC-4 Andreas Rogge wrote:
>>>
>>>> Hi James, 
>>>>
>>>> Am 13.10.25 um 17:59 schrieb James Pulver: 
>>>> > I've been trying for weeks to get this to work. I now have managed to 
>>>> > set up the dplcompat plugin, and have been able to follow 
>>>> > troubleshooting like 
>>>> https://github.com/bareos/bareos/discussions/2063 
>>>> > from the command line - CLI commands all work. But when I try and run 
>>>> a 
>>>> > backup to the s3 storage device in BAREOS it just hangs. No errors 
>>>> > noted, and status in director shows the job running forever, and 
>>>> status 
>>>> > on the storage shows 
>>>> > 
>>>> > Device status: 
>>>> > 
>>>> > Device "obj" (ObjectStorage) is being acquired with: 
>>>> >     Volume:      Full-0011 
>>>> >     Pool:        Full 
>>>> >     Media type:  obj 
>>>> > Backend connection is working. 
>>>> This means the backend connection test was successful. 
>>>> > Inflight chunks: 0 
>>>> > Pending IO flush requests: 
>>>> >    /Full-0011/0000 - 230 (try=2) 
>>>> This means chunk 0000 (i.e. the first part of the volume) should be 
>>>> written to the backend storage. Its size is 230 bytes, so probably the 
>>>> volume header. This operation failed for some reason and we are now in 
>>>> try 2. 
>>>>
>>>> > It never progresses. I'm really not sure how to proceed. 
>>>> I think the way to go here is to enable debug logging and tracing in 
>>>> the 
>>>> SD so you'll see what happens with the backend calls. 
>>>>
>>>> You can just do "setdebug trace=1 storage=your-storage level=130" in 
>>>> bconsole. It will print the path to the trace-file that will be created 
>>>> on the SD. 
>>>> In that tracefile you should be able to see what is happening between 
>>>> Bareos and your script and maybe figure out where it fails. 
>>>> You can also send it here (or to me personally, if you don't want to 
>>>> share that publicly) and I'll see if I can help. 
>>>>
>>>> Best Regards, 
>>>> Andreas 
>>>> -- 
>>>> Andreas Rogge [email protected] 
>>>> Bareos GmbH & Co. KG Phone: +49 221-630693-86 <+49%20221%2063069386> 
>>>> http://www.bareos.com 
>>>>
>>>> Sitz der Gesellschaft: Köln | Amtsgericht Köln: HRA 29646 
>>>> Komplementär: Bareos Verwaltungs-GmbH 
>>>> Geschäftsführer: Stephan Dühr, Jörg Steffens, Philipp Storz 
>>>>
>>>>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/d/msgid/bareos-users/40b6716e-c4f0-4d13-b6d2-bf5bea80e4b5n%40googlegroups.com.

Reply via email to