Hi Martin, Thanks for your help, I will contact them
Met vriendelijke groet / With kind regards, Rick > On May 14, 2020, at 6:48 PM, Martin Simmons <mar...@lispworks.com> wrote: > > Yes, there is no workaround. I suggest you contact Bacula Systems about it, > becasue they might not know about the problem. > > __Martin > > >>>>>> On Thu, 14 May 2020 16:38:06 +0000, Rick Tuk said: > >> Hi Martin, >> >> I installed Bacula through the Ubuntu repository provided by Bacula: >> >> deb http://bacula.org/packages/<key>/debs/9.6.3/bionic/amd64 bionic main >> >> Other than building bacula myself or waiting for a new release there is >> probably not much that I can do, right? >> >> Met vriendelijke groet / With kind regards, >> Rick >> >>>> On May 14, 2020, at 6:20 PM, Martin Simmons <mar...@lispworks.com> wrote: >>> >>> OK, so your bacula-sd-cloud-driver-9.6.3.so is built without S3 support for >>> some reason (there should be many lines mentioning s3_driver). The output >>> of >>> objdump is also strangely mangled, because all of the references to >>> fil_driver >>> should be file_driver. >>> >>> It looks like libs3 was not detected when Bacula was built. >>> >>> __Martin >>> >>>>>>>> On Thu, 14 May 2020 15:08:09 +0000, Rick Tuk said: >>>> >>>> Hi Martin, >>>> >>>> The Plugin Directory is /opt/bacula/plugins: >>>> >>>> Storage { >>>> Name = soteria-sd >>>> SD Address = soteria.local.domain >>>> SD Port = 9103 >>>> Maximum Concurrent Jobs = 20 >>>> >>>> Working Directory = /opt/bacula/working >>>> Pid Directory = /opt/bacula/working >>>> >>>> Plugin Directory = "/opt/bacula/plugins" >>>> } >>>> >>>> The output of objdump -t >>>> /opt/bacula/plugins/bacula-sd-cloud-driver-9.6.3.so | grep _driver is: >>>> >>>> 0000000000000000 l df *ABS* 0000000000000000 s3_driver.c >>>> 0000000000000000 l df *ABS* 0000000000000000 >>>> file_driver.c >>>> 0000000000014f60 l O .rodata 000000000000004a >>>> _ZZN11fil_driver22get_cloud_volumes_listEP3DCRP5alistRPcE19__PRETTY_FUNCTION__ >>>> 0000000000014fc0 l O .rodata 000000000000005c >>>> _ZZN11fil_driver27get_cloud_volume_parts_listEP3DCRPKcP5ilistRPcE19__PRETTY_FUNCTION__ >>>> 0000000000015100 l O .rodata 000000000000004c >>>> _ZZN11fil_driver10put_objectEP8transferPKcS3_P7bwlimitE19__PRETTY_FUNCTION__ >>>> 00000000000150a0 l O .rodata 0000000000000049 >>>> _ZZN11fil_driver19make_cloud_filenameERPcPKcjE19__PRETTY_FUNCTION__ >>>> 0000000000015060 l O .rodata 000000000000003e >>>> _ZZN11fil_driver24copy_cache_part_to_cloudEP8transferE19__PRETTY_FUNCTION__ >>>> 0000000000015020 l O .rodata 000000000000003e >>>> _ZZN11fil_driver24copy_cloud_part_to_cacheEP8transferE19__PRETTY_FUNCTION__ >>>> 0000000000012b90 g F .text 000000000000010e >>>> _ZN11file_driver24copy_cache_part_to_cloudEP8transfer >>>> 0000000000011490 g F .text 0000000000000088 >>>> _ZN11file_driver4initEP3JCRP9cloud_devP6DEVRES >>>> 0000000000012960 g F .text 000000000000022c >>>> _ZN11file_driver21truncate_cloud_volumeEP3DCRPKcP5ilistRPc >>>> 0000000000012090 g F .text 00000000000007f6 >>>> _ZN11file_driver10put_objectEP8transferPKcS3_P7bwlimit >>>> 0000000000217d20 w O .data.rel.ro 0000000000000068 >>>> _ZTV11fil_driver >>>> 0000000000011a60 g F .text 000000000000062f >>>> _ZN11file_driver27get_cloud_volume_parts_listEP3DCRPKcP5ilistRPc >>>> 000000000000f7a0 w F .text 0000000000000040 >>>> _ZN11file_driverD0Ev >>>> 000000000000f7e0 w F .text 000000000000002c >>>> _ZN11file_driverD2Ev >>>> 0000000000010950 g F .text 0000000000000185 >>>> _ZN16transfer_manager8get_xferEmPFPvP8transferEPcPKcjP12cloud_driverP3DCRP11cloud_proxy >>>> 0000000000011580 g F .text 00000000000004d5 >>>> _ZN11file_driver22get_cloud_volumes_listEP3DCRP5alistRPc >>>> 0000000000011520 g F .text 0000000000000006 >>>> _ZN11file_driver10end_of_jobEP3DCR >>>> 0000000000010250 g F .text 0000000000000154 >>>> _ZN8transferC2EmPFPvPS_EPKcS5_jP12cloud_driverP3DCRP11cloud_proxy >>>> 0000000000011520 g F .text 0000000000000006 >>>> _ZN11file_driver4termEP3DCR >>>> 0000000000011530 g F .text 0000000000000047 >>>> _ZN11file_driver12start_of_jobEP3DCR >>>> 0000000000012ca0 g F .text 00000000000000d3 >>>> _ZN11file_driver24copy_cloud_part_to_cacheEP8transfer >>>> 000000000000f7e0 w F .text 000000000000002c >>>> _ZN11file_driverD1Ev >>>> 00000000000128a0 g F .text 00000000000000bb >>>> _ZN11file_driver19make_cloud_filenameERPcPKcj >>>> 0000000000012890 g F .text 000000000000000c >>>> _ZN11file_driver16get_cloud_objectEP8transferPKcS3_ >>>> 0000000000010250 g F .text 0000000000000154 >>>> _ZN8transferC1EmPFPvPS_EPKcS5_jP12cloud_driverP3DCRP11cloud_proxy >>>> 0000000000217948 w O .data.rel.ro 0000000000000068 >>>> _ZTV12cloud_driver >>>> >>>> >>>> Met vriendelijke groet / With kind regards, >>>> Rick >>>> >>>>> On May 14, 2020, at 3:19 PM, Martin Simmons <mar...@lispworks.com> wrote: >>>>> >>>>> It looks to me like S3 support is missing. >>>>> >>>>> What is the PluginDirectory in your bacula-sd.conf? >>>>> >>>>> Find the bacula-sd-cloud-driver-9.6.3.so in that directory and post the >>>>> output >>>>> of: >>>>> >>>>> objdump -t /...path.../...to.../bacula-sd-cloud-driver-9.6.3.so | grep >>>>> _driver >>>>> >>>>> __Martin >>>>> >>>>> >>>>>>>>>> On Thu, 14 May 2020 06:24:53 +0000, Rick Tuk said: >>>>>> >>>>>> LS, >>>>>> >>>>>> I read Martin Simmons' reply to Phillip Dale’s message. >>>>>> The traceback I found did not have much information in it, so I >>>>>> installed gd and changed the btraceback script to run gd as root. >>>>>> >>>>>> The following traceback is triggered by opening bconsole and trying to >>>>>> list all volumes in cloud, same sd crash happens: >>>>>> >>>>>> [New LWP 5942] >>>>>> [New LWP 6084] >>>>>> [Thread debugging using libthread_db enabled] >>>>>> Using host libthread_db library >>>>>> "/lib/x86_64-linux-gnu/libthread_db.so.1". >>>>>> 0x00007ff1e39ca03f in select () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> $1 = "14-May-2020 08:18:11\000\000\000\000\000\000\000\000\000" >>>>>> $2 = 0x56298dfb0ee0 <my_name> "soteria.svc.mostwanted.io-sd" >>>>>> $3 = 0x56298ee4d0e8 "bacula-sd" >>>>>> $4 = 0x56298ee4d128 "/opt/bacula/bin/bacula-sd" >>>>>> $5 = 0x0 >>>>>> $6 = '\000' <repeats 49 times> >>>>>> $7 = 0x7ff1e42ad55b "9.6.3 (09 March 2020)" >>>>>> $8 = 0x7ff1e42ad53a "x86_64-pc-linux-gnu" >>>>>> $9 = 0x7ff1e42ad533 "ubuntu" >>>>>> $10 = 0x7ff1e42ad555 "18.04" >>>>>> $11 = "soteria", '\000' <repeats 42 times> >>>>>> $12 = 0x7ff1e42ad54e "ubuntu 18.04" >>>>>> Environment variable "TestName" not defined. >>>>>> #0 0x00007ff1e39ca03f in select () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> #1 0x00007ff1e4269618 in bnet_thread_server (addrs=<optimized out>, >>>>>> max_clients=41, client_wq=0x56298dfb1020 <dird_workq>, >>>>>> handle_client_request=0x56298dd99ee0 <handle_connection_request(void*)>) >>>>>> at bnet_server.c:166 >>>>>> #2 0x000056298dd9126a in main (argc=<optimized out>, argv=<optimized >>>>>> out>) at stored.c:326 >>>>>> >>>>>> Thread 3 (Thread 0x7ff1e1e6f700 (LWP 6084)): >>>>>> #0 0x00007ff1e403f23a in waitpid () from >>>>>> /lib/x86_64-linux-gnu/libpthread.so.0 >>>>>> #1 0x00007ff1e429453e in signal_handler (sig=11) at signal.c:233 >>>>>> #2 <signal handler called> >>>>>> #3 0x00007ff1e0c64787 in cloud_dev::get_cloud_volumes_list >>>>>> (this=<optimized out>, dcr=0x7ff1dc00a138, volumes=0x7ff1e1e6ec50, >>>>>> err=@0x7ff1e1e6ec48: 0x7ff1dc001330 "") at cloud_dev.h:110 >>>>>> #4 0x000056298dd95829 in cloud_list_cmd (jcr=<optimized out>) at >>>>>> dircmd.c:815 >>>>>> #5 0x000056298dd9a394 in handle_connection_request (arg=0x56298ee57428) >>>>>> at dircmd.c:242 >>>>>> #6 0x00007ff1e429f518 in workq_server (arg=0x56298dfb1020 <dird_workq>) >>>>>> at workq.c:372 >>>>>> #7 0x00007ff1e40346db in start_thread () from >>>>>> /lib/x86_64-linux-gnu/libpthread.so.0 >>>>>> #8 0x00007ff1e39d488f in clone () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> >>>>>> Thread 2 (Thread 0x7ff1e166e700 (LWP 5942)): >>>>>> #0 0x00007ff1e403af85 in pthread_cond_timedwait@@GLIBC_2.3.2 () from >>>>>> /lib/x86_64-linux-gnu/libpthread.so.0 >>>>>> #1 0x00007ff1e429eb56 in watchdog_thread (arg=<optimized out>) at >>>>>> watchdog.c:299 >>>>>> #2 0x00007ff1e40346db in start_thread () from >>>>>> /lib/x86_64-linux-gnu/libpthread.so.0 >>>>>> #3 0x00007ff1e39d488f in clone () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> >>>>>> Thread 1 (Thread 0x7ff1e4b59300 (LWP 5938)): >>>>>> #0 0x00007ff1e39ca03f in select () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> #1 0x00007ff1e4269618 in bnet_thread_server (addrs=<optimized out>, >>>>>> max_clients=41, client_wq=0x56298dfb1020 <dird_workq>, >>>>>> handle_client_request=0x56298dd99ee0 <handle_connection_request(void*)>) >>>>>> at bnet_server.c:166 >>>>>> #2 0x000056298dd9126a in main (argc=<optimized out>, argv=<optimized >>>>>> out>) at stored.c:326 >>>>>> #0 0x00007ff1e39ca03f in select () from /lib/x86_64-linux-gnu/libc.so.6 >>>>>> No symbol table info available. >>>>>> #1 0x00007ff1e4269618 in bnet_thread_server (addrs=<optimized out>, >>>>>> max_clients=41, client_wq=0x56298dfb1020 <dird_workq>, >>>>>> handle_client_request=0x56298dd99ee0 <handle_connection_request(void*)>) >>>>>> at bnet_server.c:166 >>>>>> 166 bnet_server.c: No such file or directory. >>>>>> maxfd = 7 >>>>>> sockset = {fds_bits = {128, 0 <repeats 15 times>}} >>>>>> newsockfd = <optimized out> >>>>>> stat = <optimized out> >>>>>> clilen = 16 >>>>>> clientaddr = {ss_family = 2, __ss_padding = >>>>>> "\262\354\n`\bd\000\000\000\000\000\000\000\000\000\352٠\375\177\000\000ݲ\233\343\361\177\000\000\060\346٠\375\177\000\000\036\260\224\344\361\177\000\000\220\245\265\344\361\177\000\000\060\346٠\375\177\000\000\240\361\377\177\003\000\000\000\307c\224\344\361\177\000\000`\346٠\375\177\000\000\030\352٠\375\177\000\000X\232W\342\361\177\000\000X\232!\000\000\000\000\000\000\200W\342\361\177\000", >>>>>> __ss_align = 1} >>>>>> tlog = <optimized out> >>>>>> turnon = 1 >>>>>> request = {fd = 8, user = '\000' <repeats 127 times>, daemon = >>>>>> "soteria.svc.mostwanted.io-sd", '\000' <repeats 99 times>, pid = >>>>>> "5938\000\000\000\000\000", client = {{name = '\000' <repeats 127 >>>>>> times>, addr = '\000' <repeats 127 times>, sin = 0x7ff1e38b27e0, unit = >>>>>> 0x0, request = 0x7ffda0d9e5a0}}, server = {{name = '\000' <repeats 127 >>>>>> times>, addr = '\000' <repeats 127 times>, sin = 0x7ff1e38b2760, unit = >>>>>> 0x0, request = 0x7ffda0d9e5a0}}, sink = 0x0, hostname = 0x7ff1e36aeb30 >>>>>> <sock_hostname>, hostaddr = 0x7ff1e36aeae0 <sock_hostaddr>, cleanup = >>>>>> 0x0, config = 0x0} >>>>>> addr = <optimized out> >>>>>> fd_ptr = 0x0 >>>>>> buf = "10.96.8.100", '\000' <repeats 116 times> >>>>>> sockfds = {<SMARTALLOC> = {<No data fields>}, head = 0x7ffda0d9e3d0, >>>>>> tail = 0x7ffda0d9e3d0, loffset = 0, num_items = 1} >>>>>> allbuf = >>>>>> "hM\266\344\361\177\000\000\210\227\224\344\361\177\000\000\020J\266\344\361\177\000\000\020J\266\344\361\177\000\000\000\000\000\000\000\000\000\000\320L%\344\361\177\000\000\033p\351\003\000\000\000\000$iL\344\361\177\000\000\300EL\344\361\177\000\000X\311$\344\361\177\000\000\020J\266\344\361\177\000\000\000\000\000\000\361\177\000\000p\353٠\375\177\000\000\003\000\000\000\361\177\000\000`\353٠\375\177\000\000\000\000\000\000\375\177\000\000ج\265\344\361\177\000\000\000\000\000\000\000\000\000\000\020\000\000\000\000\000\000\000\001\000\000\000\223T`\275ج\265\344\361\177\000\000\301\006\\\372\000\000\000\000\330\301\344\216)V\000\000\326}'\344\361\177\000\000\330\301\344\216)V\000\000\300\313*\344"... >>>>>> #2 0x000056298dd9126a in main (argc=<optimized out>, argv=<optimized >>>>>> out>) at stored.c:326 >>>>>> 326 stored.c: No such file or directory. >>>>>> ch = <optimized out> >>>>>> no_signals = <optimized out> >>>>>> thid = 140676853856000 >>>>>> uid = 0x0 >>>>>> gid = 0x0 >>>>>> #0 0x0000000000000000 in ?? () >>>>>> No symbol table info available. >>>>>> #0 0x0000000000000000 in ?? () >>>>>> No symbol table info available. >>>>>> #0 0x0000000000000000 in ?? () >>>>>> No symbol table info available. >>>>>> #0 0x0000000000000000 in ?? () >>>>>> No symbol table info available. >>>>>> #0 0x0000000000000000 in ?? () >>>>>> No symbol table info available. >>>>>> Attempt to dump current JCRs. njcrs=1 >>>>>> threadid=0x7ff1e1e6f700 JobId=0 JobStatus=C jcr=0x7ff1dc0008f8 >>>>>> name=*System* >>>>>> use_count=1 killable=1 >>>>>> JobType=I JobLevel= >>>>>> sched_time=14-May-2020 08:18 start_time=01-Jan-1970 01:00 >>>>>> end_time=01-Jan-1970 01:00 wait_time=01-Jan-1970 01:00 >>>>>> db=(nil) db_batch=(nil) batch_started=0 >>>>>> dcr=*None* >>>>>> List plugins. Hook count=0 >>>>>> >>>>>> Met vriendelijke groet / With kind regards, >>>>>> Rick Tuk >>>>>> >>>>>> On May 7, 2020, at 1:32 PM, Rick Tuk <r...@mostwanted.io> wrote: >>>>>> >>>>>> LS, >>>>>> >>>>>> I am trying to get Bacula 9.6.3 up and running on Ubuntu 18.04 using the >>>>>> bacula-cloud-storage package to store the backups to an Ceph cluster >>>>>> using the S3 interface. >>>>>> All services are running, when I manually try to run a backup job (in >>>>>> this case a backup of the same host) The job fails with Fatal error: >>>>>> job.c:3011 Comm error with SD. bad response to Append Data. ERR=No data >>>>>> available >>>>>> When this happens the SD daemon crashes with the following error: Bacula >>>>>> interrupted by signal 11: Segmentation violation >>>>>> >>>>>> Configs related to this setup that might be relevant: >>>>>> >>>>>> bacula-dir: >>>>>> >>>>>> Storage { >>>>>> Name = Full >>>>>> Address = soteria.local.domain >>>>>> SD Port = 9103 >>>>>> Password = “removed-for-security" >>>>>> Device = Full >>>>>> Media Type = CloudType >>>>>> } >>>>>> >>>>>> Storage { >>>>>> Name = Diff >>>>>> Address = soteria.local.domain >>>>>> SD Port = 9103 >>>>>> Password = “removed-for-security" >>>>>> Device = Diff >>>>>> Media Type = CloudType >>>>>> } >>>>>> >>>>>> Storage { >>>>>> Name = Inc >>>>>> Address = soteria.local.domain >>>>>> SD Port = 9103 >>>>>> Password = “removed-for-security" >>>>>> Device = Diff >>>>>> Media Type = CloudType >>>>>> } >>>>>> >>>>>> Pool { >>>>>> Name = Daily >>>>>> Pool Type = Backup >>>>>> Recycle = yes >>>>>> AutoPrune = yes >>>>>> Storage = Inc >>>>>> File Retention = 1 months >>>>>> Job Retention = 1 months >>>>>> Volume Retention = 1 months >>>>>> Maximum Volume Bytes = 10G >>>>>> Label Format = daily- >>>>>> } >>>>>> >>>>>> bacula-sd: >>>>>> >>>>>> Cloud { >>>>>> Name = Ceph-S3 >>>>>> Driver = "S3" >>>>>> HostName = “s3.local.domain" >>>>>> BucketName = "bacula" >>>>>> AccessKey = “removed-for-security" >>>>>> SecretKey = “removed-for-security" >>>>>> Protocol = HTTPS >>>>>> UriStyle = Path >>>>>> Truncate Cache = No >>>>>> Upload = EachPart >>>>>> } >>>>>> >>>>>> Device { >>>>>> Name = Full >>>>>> Cloud = Ceph-S3 >>>>>> Archive Device = /bacula/backup/full >>>>>> Device Type = Cloud >>>>>> Media Type = CloudType >>>>>> Maximum Part Size = 10 MB >>>>>> Label Media = yes >>>>>> Random Access = yes >>>>>> Automatic Mount = yes >>>>>> Removable Media = no >>>>>> Always Open = no >>>>>> } >>>>>> >>>>>> Device { >>>>>> Name = Diff >>>>>> Cloud = Ceph-S3 >>>>>> Archive Device = /bacula/backup/diff >>>>>> Device Type = Cloud >>>>>> Media Type = CloudType >>>>>> Maximum Part Size = 10 MB >>>>>> Label Media = yes >>>>>> Random Access = yes >>>>>> Automatic Mount = yes >>>>>> Removable Media = no >>>>>> Always Open = no >>>>>> } >>>>>> >>>>>> Device { >>>>>> Name = Inc >>>>>> Cloud = Ceph-S3 >>>>>> Archive Device = /bacula/backup/inc >>>>>> Device Type = Cloud >>>>>> Media Type = CloudType >>>>>> Maximum Part Size = 10 MB >>>>>> Label Media = yes >>>>>> Random Access = yes >>>>>> Automatic Mount = yes >>>>>> Removable Media = no >>>>>> Always Open = no >>>>>> } >>>>>> >>>>>> If any additional information is required, please let me know, I’m >>>>>> really hoping to get this to work soon. >>>>>> >>>>>> Met vriendelijke groet / With kind regards, >>>>>> >>>>>> Rick >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Bacula-users mailing list >>>>>> Bacula-users@lists.sourceforge.net >>>>>> https://lists.sourceforge.net/lists/listinfo/bacula-users >>>>>> >>>> >>>> >> >> _______________________________________________ Bacula-users mailing list Bacula-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/bacula-users