Hi, I setup ceph cluster thru mkcephfs command, after I enter "ceph -s", it always returns 4950 stuck unclean pgs. I tried the same "ceph -s" after 12 hrs, there still returns the same unclean pgs number, nothing changed. Does mkcephfs always has the problem or I did something wrong? I attached the result of "ceph -s", "ceph osd tree" and ceph.conf I have, please kindly help.
[root@ceph]# ceph -s
cluster 99fd4ff8-0fb8-47b9-8179-fefbba1c2503
health HEALTH_WARN 4950 pgs degraded; 4950 pgs stuck unclean; recovery
21/42 objects degraded (50.000%); 3/24 in osds are down; clock skew detected on
mon.1, mon.2
monmap e1: 3 mons at
{0=192.168.0.2:6789/0,1=192.168.0.3:6789/0,2=192.168.0.4:6789/0}, election
epoch 6, quorum 0,1,2 0,1,2
mdsmap e4: 1/1/1 up {0=0=up:active}
osdmap e6019: 24 osds: 21 up, 24 in
pgmap v16445: 4950 pgs, 6 pools, 9470 bytes data, 21 objects
4900 MB used, 93118 MB / 98019 MB avail
21/42 objects degraded (50.000%)
4950 active+degraded
[root@ceph]# ceph osd tree //part of returns
# id weight type name up/down reweight
-36 25 root vsm
-31 3.2 storage_group ssd
-16 3 zone zone_a_ssd
-1 1 host vsm2_ssd_zone_a
2 1 osd.2 up 1
-6 1 host vsm3_ssd_zone_a
10 1 osd.10 up 1
-11 1 host vsm4_ssd_zone_a
18 1 osd.18 up 1
-21 0.09999 zone zone_c_ssd
-26 0.09999 zone zone_b_ssd
-33 3.2 storage_group sata
-18 3 zone zone_a_sata
-3 1 host vsm2_sata_zone_a
1 1 osd.1 up 1
-8 1 host vsm3_sata_zone_a
9 1 osd.9 up 1
-13 1 host vsm4_sata_zone_a
17 1 osd.17 up 1
-23 0.09999 zone zone_c_sata
-28 0.09999 zone zone_b_sata
Wei Cao (Buddy)
ceph.conf
Description: ceph.conf
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
