Dear All

We have the same question here, if anyone can help ... Thank you!

We did not find any documentation about the steps to reset & restart the sync.
Especially the implications of 'bilog trim', 'mdlog trim' and 'datalog trim'.

Our secondary zone is read-only. Both master and secondary zone on Nautilus 
(master 14.2.9 and secondary 14.2.12).

Can someone also clarify following points? Many thanks in advance!

1) Is it safe to use these 3 commands (bilog trim, mdlog trim, datalog trim) on 
the master ?
  Are the bi logs exclusively used for the sync or are they needed even without 
multi-site? (mdlog/datalog are obviously only for multi-site)

2) Can we run these 3 commands during the sync or do we need first to stop all 
instances on the secondary zone ?
  In the latter case, do we need to stop the client traffic and wait on md/data 
sync to catch up prior to stop the secondary zone instances?

3) Can we then restart the instances on the secondary zone and expect rgw sync 
to run correctly ?
  Or do we need first to run 'metadata sync init' and 'data sync init' on the 
secondary zone ? (to trigger a full sync)
  Or is it necessary to delete all rgw pools on the secondary zone ?

4) And regarding the full sync, is it verifying the full object data, or only 
object size and mtime?
  If we update the secondary zone to Nautilus 14.2.18 and enable 
rgw_sync_obj_etag_verify,
  does a full sync will also detect ETag mismatches on objects that are already 
present on the secondary zone?


Cheers
Francois



________________________________
From: ceph-users on behalf of Osiński Piotr <[email protected]>
Sent: Saturday, June 22, 2019 11:44 AM
To: [email protected]
Subject: [ceph-users] How to reset and configure replication on multiple RGW 
servers from scratch?


Hi,

For testing purposes, I configured RGW multisite synchronization between two 
ceph mimic 13.2.6 clusters (I also tried: 13.2.5).
Now I want to reset all current settings and configure replication from scratch.

Data(pools, buckets) on the master zone will not be deleted.

What has been done:
1) Deleted the secondary zone
# radosgw-admin zone delete --rgw-zone=dc2_zone

2) Removed the secondary zone from zonegroup
# radosgw-admin zonegroup remove --rgw-zonegroup=master_zonegroup 
--rgw-zone=dc2_zone

3) Commited changes
# radosgw-admin period update --commit

4) Trimmed all datalogs on master zone
# radosgw-admin datalog trim --start-date="2019-06-12 12:01:54" 
--end-date="2019-06-22 12:01:56"

5) Trimmed all error sync on master zone
# radosgw-admin sync error trim --start-date="2019-06-07 07:19:26" 
--end-date="2019-06-22 15:59:00"

6) Deleted and recreated empty pools on secondary cluster:
dc2_zone.rgw.control
dc2_zone.rgw.meta
dc2_zone.rgw.log
dc2_zone.rgw.buckets.index
dc2_zone.rgw.buckets.data

Should I clear any other data / metadata in the master zone?
Can data be kept somewhere in the master zone that may affect the new 
replication statement?

I'm trying to track down a problem with blocked shards synchronization.


Thank you in advance for your help.


Best regards,

Piotr Osiński

<< ATT00001.txt (0.4KB) (0.4KB) >>

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to