Hi , i have an issue with my ceph cluster were two nodes wereby accident and have been recreated.
ceph osd tree # id weight type name up/down reweight -1 14.56 root default -6 14.56 datacenter dc1 -7 14.56 row row1 -9 14.56 rack rack2 -2 3.64 host ceph01 0 1.82 osd.0 up 1 1 1.82 osd.1 up 1 2 1.82 osd.2 up 1 -3 3.64 host ceph02 3 1.82 osd.3 up 1 8 1.82 osd.8 up 1 9 1.82 osd.9 up 1 -4 3.64 host ceph03 4 1.82 osd.4 up 1 5 1.82 osd.5 up 1 -5 3.64 host ceph04 6 1.82 osd.6 up 1 7 1.82 osd.7 up 1 For the past four hours the recovery process is running . I 'm not sure if it will comeback. what I have noticed is that some osd are going down during recovery and coming back. --snip-- ceph -s cluster 23d53990-4458-4faf-a598-9c60036a51f3 health HEALTH_WARN 18 pgs down; 1814 pgs peering; 1948 pgs stuck inactive; 1948 pgs stuck unclean; 1 requests are blocked > 32 sec; 1/8 in osds are down monmap e1: 3 mons at {mon01= 172.16.101.5:6789/0,mon02=172.16.101.6:6789/0,mon03=172.16.101.7:6789/0}, election epoch 46, quorum 0,1,2 mon01,mon02,mon03 osdmap e9077: 10 osds: 7 up, 8 in pgmap v127105: 2240 pgs, 7 pools, 0 bytes data, 0 objects 400M used, 14896 GB / 14896 GB avail 134 creating 1316 peering 394 creating+peering 292 active+clean 86 remapped+peering 18 down+peering --snip-- Should I wait or should I just zap and start from scratch, I dont have any data on my ceph cluster ?
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com