Hi, I have a situation where I moved the interfaces over which my ceph-public
network is connected (only the interfaces, not the IPs, etc.) this was done to
increase available bandwidth, but it backfired catastrophically. My monitors
all failed and somehow became corrupted, but I was unable to repair them. So I
rebuild the monitors in hope that I could add the existing OSDs back in and
recover the cluster.
There are three hosts. Each has a monitor and 6 osds. Each osd is a spinning
disk partition with a journal located on a SSD partition on the same host. From
what I can tell, all the data on the osd disks is intact, but even after (what
I think was) adding all the OSDs back into the crushmap, etc. the cluster
doesn't seem like it is "seeing" the partitions and I'm at a loss for how to
troubleshoot it further.
Hosts are all Ubutunu trusty running 0.80.7 ceph packages.
dgeist# ceph -s
cluster ac486394-802a-49d3-a92c-a103268ea189
health HEALTH_WARN 4288 pgs stuck inactive; 4288 pgs stuck unclean; 18/18
in osds are down
monmap e1: 3 mons at
{hypd01=10.100.100.11:6789/0,hypd02=10.100.100.12:6789/0,hypd03=10.100.100.13:6789/0},
election epoch 40, quorum 0,1,2 hypd01,hypd02,hypd03
osdmap e65: 18 osds: 0 up, 18 in
pgmap v66: 4288 pgs, 4 pools, 0 bytes data, 0 objects
0 kB used, 0 kB / 0 kB avail
4288 creating
dgeist# ceph osd tree
# id weight type name up/down reweight
-1 18 root default
-2 6 host hypd01
0 1 osd.0 down 1
1 1 osd.1 down 1
2 1 osd.2 down 1
3 1 osd.3 down 1
4 1 osd.4 down 1
5 1 osd.5 down 1
-3 6 host hypd02
6 1 osd.6 down 1
7 1 osd.7 down 1
8 1 osd.8 down 1
9 1 osd.9 down 1
10 1 osd.10 down 1
11 1 osd.11 down 1
-4 6 host hypd03
12 1 osd.12 down 1
13 1 osd.13 down 1
14 1 osd.14 down 1
15 1 osd.15 down 1
16 1 osd.16 down 1
17 1 osd.17 down 1
Thanks in advance for any thoughts on how to recover this.
Dan
Dan Geist dan(@)polter.net
(33.942973, -84.312472)
http://www.polter.net
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com