Re: [ceph-users] Object Gateway Cloud Sync to S3

2019-02-05 Thread Ryan
On Tue, Feb 5, 2019 at 3:35 PM Ryan  wrote:

> I've been trying to configure the cloud sync module to push changes to an
> Amazon S3 bucket without success. I've configured the module according to
> the docs with the trivial configuration settings. Is there an error log I
> should be checking? Is the "radosgw-admin sync status
> --rgw-zone=mycloudtierzone" the correct command to check status?
>
> Thanks,
> Ryan
>

It turns out I can get it to sync as long as I leave "radosgw-admin
--rgw-zone=aws-docindex data sync run --source-zone=default" running. I
thought with mimic the sync was built into the ceph-radosgw service? I'm
running version 13.2.4. I'm also seeing these errors on the console while
running that command.

2019-02-05 17:40:10.679 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.694 7fb1ef06b680  0 RGW-SYNC:data:sync:shard[25]:
ERROR: failed to read remote data log info: ret=-2
2019-02-05 17:40:10.695 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.711 7fb1ef06b680  0 RGW-SYNC:data:sync:shard[43]:
ERROR: failed to read remote data log info: ret=-2
2019-02-05 17:40:10.712 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2
2019-02-05 17:40:10.720 7fb1ef06b680  0 meta sync: ERROR:
RGWBackoffControlCR called coroutine returned -2

Additionally "radosgw-admin --rgw-zone=aws-docindex data sync error list
--source-zone=default" is showing numerous error code 39 responses/

 "message": "failed to sync bucket instance: (39) Directory not empty"
"message": "failed to sync object(39) Directory not empty"

When it successfully completes I see the following

  metadata sync syncing
full sync: 0/64 shards
incremental sync: 64/64 shards
metadata is caught up with master
  data sync source: af57fe9a-43a7-4998-9574-4016f5fa6661 (default)
syncing
full sync: 0/128 shards
incremental sync: 128/128 shards
data is caught up with source

When I stop the "data sync run" the status will just sit on

  data sync source: af57fe9a-43a7-4998-9574-4016f5fa6661 (default)
syncing
full sync: 0/128 shards
incremental sync: 128/128 shards
data is behind on 1 shards
behind shards: [75]
oldest incremental change not applied: 2019-02-05
17:44:51.0.367478s

Thanks,
Ryan
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Object Gateway Cloud Sync to S3

2019-02-05 Thread Ryan
I've been trying to configure the cloud sync module to push changes to an
Amazon S3 bucket without success. I've configured the module according to
the docs with the trivial configuration settings. Is there an error log I
should be checking? Is the "radosgw-admin sync status
--rgw-zone=mycloudtierzone" the correct command to check status?

Thanks,
Ryan
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com