Hello,

My Ceph cluster with 3 nodes is showing a HEALTH_ERR, with the following errors:

 * OSD_SCRUB_ERRORS: 6 scrub errors
 * PG_DAMAGED: Possible data damage: 6 pgs inconsistent
 * CEPHADM_FAILED_DAEMON: 3 failed cephadm daemon(s)
 * MON_CLOCK_SKEW: clock skew detected on mon.ceph3
 * MON_DOWN: 1/3 mons down, quorum ceph2,ceph3
 * OSD_NEARFULL: 4 nearfull osd(s)
 * PG_NOT_DEEP_SCRUBBED: 2 pgs not deep-scrubbed in time

All OSDs (18) are up though,and I don't see any error in each server's dmesg logs for hard drive issues.

Current cluster's status page is showing: Scrubbing: Active

Is the problem recoverable?


Thanks,

Mike

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to