both cluster are in the same epoch and period:
root@ceph-a-1:~# radosgw-admin period get-current
{
"current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
}
root@ceph-b-1:~# radosgw-admin period get-current
{
"current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
}
but the sync state is still "master is on a different period":
root@ceph-b-1:~# radosgw-admin sync status
realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
metadata sync syncing
full sync: 0/64 shards
master is on a different period:
master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569
local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4
incremental sync: 64/64 shards
metadata is caught up with master
data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
syncing
full sync: 0/128 shards
incremental sync: 128/128 shards
data is caught up with source
Am Dienstag, den 14.11.2017, 18:21 +0100 schrieb Kim-Norman Sahm:
> both cluster are in the same epoch and period:
>
> root@ceph-a-1:~# radosgw-admin period get-current
> {
> "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
> }
>
> root@ceph-b-1:~# radosgw-admin period get-current
> {
> "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
> }
>
> Am Dienstag, den 14.11.2017, 17:05 +0000 schrieb David Turner:
> >
> > I'm assuming you've looked at the period in both places `radosgw-
> > admin period get` and confirmed that the second site is behind the
> > master site (based on epochs). I'm also assuming (since you linked
> > the instructions) that you've done `radosgw-admin period pull` on
> > the
> > second site to get any period updates that have been done to the
> > master site.
> >
> > If my assumptions are wrong. Then you should do those things. If
> > my
> > assumptions are correct, then running `radosgw-admin period update
> > --
> > commit` on the the master site and `radosgw-admin period pull` on
> > the
> > second site might fix this. If you've already done that as well
> > (as
> > they're steps in the article you linked), then you need someone
> > smarter than I am to chime in.
> >
> > On Tue, Nov 14, 2017 at 11:35 AM Kim-Norman Sahm <[email protected]
> > e>
> > wrote:
> > >
> > > hi,
> > >
> > > i've installed a ceph multi site setup with two ceph clusters and
> > > each
> > > one radosgw.
> > > the multi site setup was in sync, so i tried a failover.
> > > cluster A is going down and i've changed the zone (b) on cluster
> > > b
> > > to
> > > the new master zone.
> > > it's working fine.
> > >
> > > now i start the cluster A and try to switch back the master zone
> > > to
> > > A.
> > > cluster A believes that he is the master, cluster b is secondary.
> > > but on the secondary is a different period and the bucket delta
> > > is
> > > not
> > > synced to the new master zone:
> > >
> > > root@ceph-a-1:~# radosgw-admin sync status
> > > realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
> > > zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
> > > zone 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
> > > metadata sync no sync (zone is master)
> > > data sync source: 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
> > > syncing
> > > full sync: 0/128 shards
> > > incremental sync: 128/128 shards
> > > data is caught up with source
> > >
> > > root@ceph-b-1:~# radosgw-admin sync status
> > > realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
> > > zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
> > > zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
> > > metadata sync syncing
> > > full sync: 0/64 shards
> > > master is on a different period:
> > > master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569
> > > local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4
> > > incremental sync: 64/64 shards
> > > metadata is caught up with master
> > > data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
> > > syncing
> > > full sync: 0/128 shards
> > > incremental sync: 128/128 shards
> > > data is caught up with source
> > >
> > > how can i force sync the period and the bucket deltas?
> > > i've used this howto: http://docs.ceph.com/docs/master/radosgw/mu
> > > lt
> > > isit
> > > e/
> > >
> > > br Kim
> > > _______________________________________________
> > > ceph-users mailing list
> > > [email protected]
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com