Hi all,

I'm upgrading ceph cluster from Hammer 0.94.9 to jewel 10.2.6.

The ceph cluster has 3 servers (one mon and one mds each) and another 6 servers with
12 osds each.
The monitoring and mds have been succesfully upgraded to latest jewel release, however
after upgrade the first osd server(12 osds), ceph is not aware of them and
are marked as down

ceph -s

 cluster 4a158d27-f750-41d5-9e7f-26ce4c9d2d45
     health HEALTH_WARN
[...]
            12/72 in osds are down
            noout flag(s) set
     osdmap e14010: 72 osds: 60 up, 72 in; 14641 remapped pgs
            flags noout
[...]

ceph osd tree

3   3.64000         osd.3          down  1.00000 1.00000
 8   3.64000         osd.8          down  1.00000 1.00000
14   3.64000         osd.14         down  1.00000 1.00000
18   3.64000         osd.18         down  1.00000          1.00000
21   3.64000         osd.21         down  1.00000          1.00000
28   3.64000         osd.28         down  1.00000          1.00000
31   3.64000         osd.31         down  1.00000          1.00000
37   3.64000         osd.37         down  1.00000          1.00000
42   3.64000         osd.42         down  1.00000          1.00000
47   3.64000         osd.47         down  1.00000          1.00000
51   3.64000         osd.51         down  1.00000          1.00000
56   3.64000         osd.56         down  1.00000          1.00000

If I run this command with one of the down osd
ceph osd in 14
osd.14 is already in.
however ceph doesn't mark it as up and the cluster health remains
in degraded state.

Do I have to upgrade all the osds to jewel first?
Any help as I'm running out of ideas?

Thanks
Jaime

--

Jaime Ibar
High Performance & Research Computing, IS Services
Lloyd Building, Trinity College Dublin, Dublin 2, Ireland.
http://www.tchpc.tcd.ie/ | ja...@tchpc.tcd.ie
Tel: +353-1-896-3725

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to