Hello, Some osd's not getting activated after a reboot operation which cause that particular osd's landing in failed state.
Here you can see mount points were not getting updated to osd-num and mounted as a incorrect mount point, which caused osd.<num> can't able to mount/activate the osd's. Env:- RHEL 7.2 - EC 4+1, v11.2.0 bluestore. #grep mnt proc/mounts /dev/sdh1 /var/lib/ceph/tmp/mnt.om4Lbq xfs rw,noatime,attr2,inode64,sunit=512,swidth=512,noquota 0 0 /dev/sdh1 /var/lib/ceph/tmp/mnt.EayTmL xfs rw,noatime,attr2,inode64,sunit=512,swidth=512,noquota 0 0 >From /var/log/messages.. -- May 26 15:39:58 cn1 systemd: Starting Ceph disk activation: /dev/sdh2... May 26 15:39:58 cn1 systemd: Starting Ceph disk activation: /dev/sdh1... May 26 15:39:58 cn1 systemd: *start request repeated too quickly for* [email protected] => suspecting this could be root cause. May 26 15:39:58 cn1 systemd: Failed to start Ceph disk activation: /dev/sdh2. May 26 15:39:58 cn1 systemd: Unit [email protected] entered failed state. May 26 15:39:58 cn1 systemd: [email protected] failed. May 26 15:39:58 cn1 systemd: start request repeated too quickly for [email protected] May 26 15:39:58 cn1 systemd: Failed to start Ceph disk activation: /dev/sdh1. May 26 15:39:58 cn1 systemd: Unit [email protected] entered failed state. May 26 15:39:58 cn1 systemd: [email protected] failed. -- But this issue will occur intermittently after a reboot operation. Note;- We haven't face this problem in Jewel. Awaiting for comments. Thanks Jayaram
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
