Hi - I have a 3 node cluster, and ran the upgrade to 16.2.6 yesterday.
All looked like it was going well, but the MDS servers are not coming up
Ceph status shows 2 failed daemons and 3 standby.
ceph status
cluster:
id: fe3a7cb0-69ca-11eb-8d45-c86000d08867
health: HEALTH_ERR
client is using insecure global_id reclaim
failed to probe daemons or devices
1 filesystem is degraded
1 filesystem has a failed mds daemon
1 filesystem is offline
1 filesystem is online with fewer MDS than max_mds
services:
mon: 2 daemons, quorum rhel1,story (age 81m)
mgr: cube.snthzq(active, since 63s), standbys: story.gffann, rhel1.cmxwxg
mds: 0/2 daemons up (2 failed), 3 standby
osd: 12 osds: 12 up (since 80m), 12 in (since 26h); 41 remapped pgs
rgw: 3 daemons active (3 hosts, 1 zones)
data:
volumes: 0/1 healthy, 1 failed
pools: 11 pools, 497 pgs
objects: 2.06M objects, 4.1 TiB
usage: 13 TiB used, 25 TiB / 38 TiB avail
pgs: 354383/6190401 objects misplaced (5.725%)
456 active+clean
35 active+remapped+backfill_wait
6 active+remapped+backfilling
io:
recovery: 36 MiB/s, 17 objects/s
And MDS metadata shows that there are only 3
ceph mds metadata
[
{
"name": home.story.rqrdtz<http://home.story.rqrdtz>,
"addr":
"[v2:192.168.2.199:6800/1255725176,v1:192.168.2.199:6801/1255725176]",
"arch": "x86_64",
"ceph_release": "pacific",
"ceph_version": "ceph version 16.2.6
(ee28fb57e47e9f88813e24bbf4c14496ca299d31) pacific (stable)",
"ceph_version_short": "16.2.6",
"container_hostname": "story.robeckert.us",
"container_image":
quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac<mailto:quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac>,
"cpu": "Intel(R) Pentium(R) Silver J5005 CPU @ 1.50GHz",
"distro": "centos",
"distro_description": "CentOS Linux 8",
"distro_version": "8",
"hostname": "story.robeckert.us",
"kernel_description": "#1 SMP Mon Jul 26 08:06:24 EDT 2021",
"kernel_version": "4.18.0-305.12.1.el8_4.x86_64",
"mem_swap_kb": "8093692",
"mem_total_kb": "32367924",
"os": "Linux"
},
{
"name": home.rhel1.ffrufi<http://home.rhel1.ffrufi>,
"addr":
"[v2:192.168.2.141:6800/169048976,v1:192.168.2.141:6801/169048976]",
"arch": "x86_64",
"ceph_release": "pacific",
"ceph_version": "ceph version 16.2.6
(ee28fb57e47e9f88813e24bbf4c14496ca299d31) pacific (stable)",
"ceph_version_short": "16.2.6",
"container_hostname": "rhel1.robeckert.us",
"container_image":
quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac<mailto:quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac>,
"cpu": "Intel(R) Core(TM) i5-2500K CPU @ 3.30GHz",
"distro": "centos",
"distro_description": "CentOS Linux 8",
"distro_version": "8",
"hostname": "rhel1.robeckert.us",
"kernel_description": "#1 SMP Mon Jul 26 08:06:24 EDT 2021",
"kernel_version": "4.18.0-305.12.1.el8_4.x86_64",
"mem_swap_kb": "12378108",
"mem_total_kb": "24408040",
"os": "Linux"
},
{
"name": home.cube.cfrali<http://home.cube.cfrali>,
"addr":
"[v2:192.168.2.142:6800/2860921355,v1:192.168.2.142:6801/2860921355]",
"arch": "x86_64",
"ceph_release": "pacific",
"ceph_version": "ceph version 16.2.6
(ee28fb57e47e9f88813e24bbf4c14496ca299d31) pacific (stable)",
"ceph_version_short": "16.2.6",
"container_hostname": "cube.robeckert.us",
"container_image":
quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac<mailto:quay.io/ceph/ceph@sha256:5d042251e1faa1408663508099cf97b256364300365d403ca5563a518060abac>,
"cpu": "AMD Ryzen 5 3600 6-Core Processor",
"distro": "centos",
"distro_description": "CentOS Linux 8",
"distro_version": "8",
"hostname": "cube.robeckert.us",
"kernel_description": "#1 SMP Mon Jul 26 08:06:24 EDT 2021",
"kernel_version": "4.18.0-305.12.1.el8_4.x86_64",
"mem_swap_kb": "0",
"mem_total_kb": "65595656",
"os": "Linux"
}
]
How do I find the ghost services and remove them, or at least force them to use
the actual MDS servers?
Thanks,
Rob
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]