I recently started with Ceph, and overall had very few issues. However during
a process of Cluster creation I must have done something wrong which created
orphan PG groups. I suspect it was broken when I removed OSD right after
initial creation, but I am guessing.
Currently here is the Ceph output:
#ceph -s
cluster 30266c5f-5e10-4027-936c-e4409667b409
health HEALTH_WARN
65 pgs stale
22 pgs stuck inactive
65 pgs stuck stale
22 pgs stuck unclean
too many PGs per OSD (456 > max 300)
monmap e7: 7 mons at
<servers>
election epoch 156, quorum 0,1,2,3,4,5,6 <servers>
mdsmap e5: 1/1/1 up {0=<server>=up:active}
osdmap e579: 14 osds: 14 up, 14 in
pgmap v1229872: 2152 pgs, 10 pools, 193 GB data, 54315 objects
661 GB used, 7299 GB / 8319 GB avail
2065 active+clean
65 stale+active+clean
22 creating
client io 287 kB/s wr, 108 op/s
One of the Stuck Inactive is 0.21 and here is the output of ceph pg map
#ceph pg map 0.21
osdmap e579 pg 0.21 (0.21) -> up [] acting []
#ceph pg dump_stuck stale
ok
pg_stat state up up_primary acting acting_primary
0.22 stale+active+clean [5,1,6] 5 [5,1,6] 5
0.1f stale+active+clean [2,0,4] 2 [2,0,4] 2
<reducted for ease of reading>
# ceph osd stat
osdmap e579: 14 osds: 14 up, 14 in
If I do
#ceph pg 0.21 query
The command freezes and never returns any output.
I suspect that the problem is that these PGs were created but the OSD that they
were initially created under disappeared. So I believe that I should just
remove these PGs, but honestly I don’t see how.
Does anybody have any ideas as to what to do next?
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com