Hi,

We added a new host to cluster and it was rebalancing.
And one PG became "inactive, peering" for very long time which created lot
of slow requests and poor performance to the whole cluster.

When I queried that PG, it showed this :

"recovery_state": [
        {
            "name": "Started/Primary/Peering/GetMissing",
            "enter_time": "2018-01-22 18:40:04.777654",
            "peer_missing_requested": [
                {
                    "osd": "77(7)",

So I assumed it was stuck getting information from osd77 and so I marked
osd.77 down.
The status of the PG changed to "active+undersized+degraded" and PG became
active again.

Can anyone know why this happened ?
If I start osd.77,again the PG becomes inactive and peering state.


Is it becase osd.77 is bad ? Or will the same happen when the PG tries to
peer again with another disk?


Any help is really appreciated

Karun
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to