Hi,
I switch some disks from manual format to ceph-deploy (because slightly
different xfs-parameters) - all disks are on a single node of an 4-node
cluster.
After rebuilding the osd-disk one PG are incomplete:
ceph -s
cluster 591db070-15c1-4c7a-b107-67717bdb87d9
health HEALTH_WARN 1 pgs incomplete; 1 pgs stuck inactive; 1 pgs
stuck unclean
monmap e7: 3 mons at
{a=172.20.2.11:6789/0,b=172.20.2.64:6789/0,c=172.20.2.65:6789/0},
election epoch 1178, quorum 0,1,2 a,b,c
mdsmap e409: 1/1/1 up {0=b=up:active}, 2 up:standby
osdmap e22002: 52 osds: 52 up, 52 in
pgmap v10177038: 7408 pgs, 5 pools, 58618 GB data, 14662 kobjects
114 TB used, 76319 GB / 189 TB avail
7405 active+clean
1 incomplete
2 active+clean+scrubbing+deep
The pg are on one of the "rebuilded" disk (osd.42):
ceph pg map 6.289
osdmap e22002 pg 6.289 (6.289) -> up [42,31] acting [42,31]
ls -lsa /var/lib/ceph/osd/ceph-42/current/6.289_head/
insgesamt 16
0 drwxr-xr-x 2 root root 6 Feb 15 20:11 .
16 drwxr-xr-x 411 root root 12288 Feb 16 03:09 ..
ls -lsa
/var/lib/ceph/osd/ceph-31/current/6.289*/
/var/lib/ceph/osd/ceph-31/current/6.289_head/:
insgesamt
20520
8 drwxr-xr-x 2 root root 4096 Feb 15 10:24
.
12 drwxr-xr-x 320 root root 8192 Feb 15 21:11
..
4100 -rw-r--r-- 1 root root 4194304 Feb 15 10:24
benchmark\udata\uproxmox4\u638085\uobject2844__head_4F14E289__6
4100 -rw-r--r-- 1 root root 4194304 Feb 15 10:24
benchmark\udata\uproxmox4\u638085\uobject3975__head_A7EBCA89__6
4100 -rw-r--r-- 1 root root 4194304 Feb 15 10:24
benchmark\udata\uproxmox4\u638085\uobject4003__head_537FE289__6
4100 -rw-r--r-- 1 root root 4194304 Feb 15 10:24
benchmark\udata\uproxmox4\u673679\uobject344__head_FF4A1289__6
4100 -rw-r--r-- 1 root root 4194304 Feb 15 10:24
benchmark\udata\uproxmox4\u673679\uobject474__head_5FC3EA89__6
/var/lib/ceph/osd/ceph-31/current/6.289_TEMP/:
insgesamt 16
4 drwxr-xr-x 2 root root 6 Feb 15 10:24 .
12 drwxr-xr-x 320 root root 8192 Feb 15 21:11 ..
How to say ceph, that the content on osd.31 is the right one?
I have tried an "ceph osd repair osd.42" without luck.
In the manual I saw only "ceph osd lost NN" but then all other data will
also rebuild to other disks I guess.
If "osd lost" the only option, how reuse osd-42? Waiting for an healthy
cluster and then recreate the disk?
Hope for an hint.
Best regards
Udo
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com