I managed to remove the warning reweighting the crashed OSD:
ceph osd crush reweight osd.33 0.8
After the recovery, the cluster is not showing the warning any more
On 29/11/16 11:18, Xabier Elkano wrote:
> Hi all,
> my cluster is in WARN state because apparently there are some pgs
> unfound. I think that I reached this situation because the metadata
> pool, this pool was in default root but without any use because I don't
> use cephfs, I only use rbd for VMs. I don't have OSDs in the default
> root, they are assigned to different roots depending on its disk type.
> My pools have specific crush rules to use the different roots. Again,
> all my pools, but not the metadata pool, it was assigned to the default
> With this situation I got problems in one OSD (I had to reset it from
> scratch) and when I restored the situation I got some pgs unfound
> because they where in the faulty OSD and belonged to the metadata pool
> with size 1. Because I didn't care of any data in the metadata pool I
> created the unfounds pgs again with "ceph pg force_create_pg 1.25".
> Finally I set a crush rule to the metadata pool to change its location
> and the pgs were created.
> But now, the cluster is showing 29 unfound pgs, but without saying what pgs.
> How can I recover from this situation? Can I remove metadata pool and
> recreate it again?
> # ceph status
> cluster 72a4a18b-ec5c-454d-9135-04362c97c307
> health HEALTH_WARN
> recovery 29/2748828 unfound (0.001%)
> monmap e13: 5 mons at
> election epoch 99672, quorum 0,1,2,3,4 mon1,mon2,mon2,mon3,mon4
> mdsmap e35323: 0/0/1 up
> osdmap e49648: 38 osds: 38 up, 38 in
> pgmap v76150847: 3065 pgs, 21 pools, 10654 GB data, 2684 kobjects
> 31111 GB used, 25423 GB / 56534 GB avail
> 29/2748828 unfound (0.001%)
> 3063 active+clean
> 2 active+clean+scrubbing
> client io 4431 kB/s rd, 15897 kB/s wr, 2385 op/s
> # ceph health detail
> HEALTH_WARN recovery 29/2748829 unfound (0.001%)
> recovery 29/2748829 unfound (0.001%)
> My cluster runs hammer 0.94.9
> 5 Servers with 7 OSDs each on Ubuntu 14.04
> 5 monitor servers.
> Thanks and Best regards,
> ceph-users mailing list
ceph-users mailing list