On Tue, Feb 5, 2019 at 3:35 PM Ryan <rswago...@gmail.com> wrote:

> I've been trying to configure the cloud sync module to push changes to an
> Amazon S3 bucket without success. I've configured the module according to
> the docs with the trivial configuration settings. Is there an error log I
> should be checking? Is the "radosgw-admin sync status
> --rgw-zone=mycloudtierzone" the correct command to check status?
>
> Thanks,
> Ryan
>

It turns out I can get it to sync as long as I leave "radosgw-admin
--rgw-zone=aws-docindex data sync run --source-zone=default" running. I
thought with mimic the sync was built into the ceph-radosgw service? I'm
running version 13.2.4. I'm also seeing these errors on the console while
running that command.

2019-02-05 17:40:10.679 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.694 7fb1ef06b680  0 RGW-SYNC:data:sync:shard[25]:
ERROR: failed to read remote data log info: ret=-2
2019-02-05 17:40:10.695 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.711 7fb1ef06b680  0 RGW-SYNC:data:sync:shard[43]:
ERROR: failed to read remote data log info: ret=-2
2019-02-05 17:40:10.712 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.720 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2

Additionally "radosgw-admin --rgw-zone=aws-docindex data sync error list
--source-zone=default" is showing numerous error code 39 responses/

 "message": "failed to sync bucket instance: (39) Directory not empty"
"message": "failed to sync object(39) Directory not empty"

When it successfully completes I see the following

  metadata sync syncing
                full sync: 0/64 shards
                incremental sync: 64/64 shards
                metadata is caught up with master
      data sync source: af57fe9a-43a7-4998-9574-4016f5fa6661 (default)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is caught up with source

When I stop the "data sync run" the status will just sit on

      data sync source: af57fe9a-43a7-4998-9574-4016f5fa6661 (default)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is behind on 1 shards
                        behind shards: [75]
                        oldest incremental change not applied: 2019-02-05
17:44:51.0.367478s

Thanks,
Ryan
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to