Hello Cephers
I am struggling with my ceph cluster health , PGS are not getting clean , i
waited for recovery process to get end was hoping after recovery PG will become
clean , but it didn’t. Can you please share your suggestions.
cluster 0ff473d9-0670-42a3-89ff-81bbfb2e676a
health HEALTH_WARN 119 pgs down; 303 pgs peering; 303 pgs stuck inactive;
303 pgs stuck unclean; mds cluster is degraded; crush map has no
n-optimal tunables
monmap e3: 3 mons at
{ceph-mon1=192.168.1.38:6789/0,ceph-mon2=192.168.1.33:6789/0,ceph-mon3=192.168.1.31:6789/0},
election epoch 4226, quo
rum 0,1,2 ceph-mon1,ceph-mon2,ceph-mon3
mdsmap e8465: 1/1/1 up {0=ceph-mon1=up:replay}
osdmap e250466: 10 osds: 10 up, 10 in
pgmap v585809: 576 pgs, 6 pools, 101933 MB data, 25453 objects
343 GB used, 5423 GB / 5767 GB avail
273 active+clean
108 peering
119 down+peering
76 remapped+peering
# id weight type name up/down reweight
-1 5.65 root default
-2 0 host ceph-node1
-3 1.72 host ceph-node2
4 0.43 osd.4 up 1
5 0.43 osd.5 up 1
6 0.43 osd.6 up 1
7 0.43 osd.7 up 1
-4 1.31 host ceph-node4
8 0.88 osd.8 up 1
1 0.43 osd.1 up 1
-5 1.31 host ceph-node5
9 0.88 osd.9 up 1
2 0.43 osd.2 up 1
-6 0.88 host ceph-node6
10 0.88 osd.10 up 1
-7 0.43 host ceph-node3
0 0.43 osd.0 up 1
Regards
karan
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com