I recently started with Ceph, and overall had very few issues.  However during 
a process of Cluster creation I must have done something wrong which created 
orphan PG groups.  I suspect it was broken when I removed OSD right after 
initial creation, but I am guessing.

Currently here is the Ceph output:
#ceph -s
    cluster 30266c5f-5e10-4027-936c-e4409667b409
     health HEALTH_WARN
            65 pgs stale
            22 pgs stuck inactive
            65 pgs stuck stale
            22 pgs stuck unclean
            too many PGs per OSD (456 > max 300)
     monmap e7: 7 mons at 
<servers>
            election epoch 156, quorum 0,1,2,3,4,5,6 <servers>
     mdsmap e5: 1/1/1 up {0=<server>=up:active}
     osdmap e579: 14 osds: 14 up, 14 in
      pgmap v1229872: 2152 pgs, 10 pools, 193 GB data, 54315 objects
            661 GB used, 7299 GB / 8319 GB avail
                2065 active+clean
                  65 stale+active+clean
                  22 creating
  client io 287 kB/s wr, 108 op/s


One of the Stuck Inactive is 0.21 and here is the output of ceph pg map

#ceph pg map 0.21
osdmap e579 pg 0.21 (0.21) -> up [] acting []

#ceph pg dump_stuck stale
ok
pg_stat state   up      up_primary      acting  acting_primary
0.22    stale+active+clean      [5,1,6] 5       [5,1,6] 5
0.1f    stale+active+clean      [2,0,4] 2       [2,0,4] 2
<reducted for ease of reading>

# ceph osd stat
     osdmap e579: 14 osds: 14 up, 14 in

If I do 
#ceph pg 0.21 query

The command freezes and never returns any output.

I suspect that the problem is that these PGs were created but the OSD that they 
were initially created under disappeared.  So I believe that I should just 
remove these PGs, but honestly I don’t see how.

Does anybody have any ideas as to what to do next?


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to