Dear Ceph Experts,
I have recently deleted a very big directory on my cephfs and a few minutes
after my dashboard start yelling :
Overall status: HEALTH_ERR
MDS_DAMAGE: 1 MDSs report damaged metadata
So I immediately log in my ceph admin node than do a ceph -s:
cluster:
id: 472dfc88-84dc-4284-a1cf-0810ea45ae19
health: HEALTH_ERR
1 MDSs report damaged metadata
services:
mon: 3 daemons, quorum ceph-n1,ceph-n2,ceph-n3
mgr: ceph-admin(active), standbys: ceph-n1
mds: cephfs-2/2/2 up {0=ceph-admin=up:active,1=ceph-n1=up:active}, 1
up:standby
osd: 17 osds: 17 up, 17 in
rgw: 1 daemon active
data:
pools: 9 pools, 1584 pgs
objects: 1093 objects, 418 MB
usage: 2765 MB used, 6797 GB / 6799 GB avail
pgs: 1584 active+clean
io:
client: 35757 B/s rd, 0 B/s wr, 34 op/s rd, 23 op/s wr
and after a few research I tried : #ceph tell mds.0 damage ls :
"damage_type": "backtrace",
"id": 2744661796,
"ino": 1099512314364,
"path": "/M3/sogetel.net/t/te/testmda3/Maildir/dovecot.index.log.2"
And so I tried to do what I saw at
https://www.mail-archive.com/[email protected]/msg35682.html
But it did not work so now I don't know how to fix it.
Can you help me ?
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com