Hi,
I recently upgraded to jewel (10.2.2) and now I'm confronted with a rather
strange behavior: recovey does not progress in the way it should. If I
restart the osds on a host, it'll get a bit better (or worse), like this:
50 pgs undersized
recovery 43775/7057285 objects degraded (0.620%)
recovery 87980/7057285 objects misplaced (1.247%)
[restart osds on node1]
44 pgs undersized
recovery 39623/7061519 objects degraded (0.561%)
recovery 92142/7061519 objects misplaced (1.305%)
[restart osds on node1]
43 pgs undersized
1116 requests are blocked > 32 sec
recovery 38181/7061529 objects degraded (0.541%)
recovery 90617/7061529 objects misplaced (1.283%)
...
The current state is this:
osdmap e38804: 53 osds: 51 up, 51 in; 66 remapped pgs
pgmap v14797137: 4388 pgs, 8 pools, 13626 GB data, 3434 kobjects
27474 GB used, 22856 GB / 50330 GB avail
38172/7061565 objects degraded (0.541%)
90617/7061565 objects misplaced (1.283%)
8/3517300 unfound (0.000%)
4202 active+clean
109 active+recovery_wait+degraded
38 active+undersized+degraded+remapped+wait_backfill
15 active+remapped+wait_backfill
11 active+clean+inconsistent
8 active+recovery_wait+degraded+remapped
3 active+recovering+undersized+degraded+remapped
2 active+recovery_wait+undersized+degraded+remapped
All the pools have size=2 min_size=1.
(All the unfound blocks are on undersized pgs, and I cannot seem to be
able to fix them without having replicas (?). They exist, but are
outdated, from an earlier problem.)
Matyas
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com