On 28. aug. 2017 08:01, hjcho616 wrote:
Hello!

I've been using ceph for long time mostly for network CephFS storage, even before Argonaut release! It's been working very well for me. Yes, I had some power outtages before and asked few questions on this list before and got resolved happily! Thank you all!

Not sure why but we've been having quite a bit of power outages lately. Ceph appear to be running OK with those going on.. so I was pretty happy and didn't thought much of it... till yesterday, When I started to move some videos to cephfs, ceph decided that it was full although df showed only 54% utilization! Then I looked up, some of the osds were down! (only 3 at that point!)

I am running pretty simple ceph configuration... I have one machine running MDS and mon named MDS1. Two OSD machines with 5 2TB HDDs and 1 SSD for journal named OSD1 and OSD2.

At the time, I was running jewel 10.2.2. I looked at some of downed OSD's log file and googled some of them... they appeared to be tied to version 10.2.2. So I just upgraded all to 10.2.9. Well that didn't solve my problems.. =P While looking at some of this.. there was another power outage! D'oh! I may need to invest in a UPS or something... Until this happened, all of the osd down were from OSD2. But OSD1 took a hit! Couldn't boot, because osd-0 was damaged... I tried xfs_repair -L /dev/sdb1 as suggested by command line.. I was able to mount it again, phew, reboot... then /dev/sdb1 is no longer accessible! Noooo!!!

So this is what I have today! I am a bit concerned as half of the osds are down! and osd.0 doesn't look good at all...
# ceph osd tree
ID WEIGHT   TYPE NAME     UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 16.24478 root default
-2  8.12239     host OSD1
  1  1.95250         osd.1      up  1.00000          1.00000
  0  1.95250         osd.0    down        0          1.00000
  7  0.31239         osd.7      up  1.00000          1.00000
  6  1.95250         osd.6      up  1.00000          1.00000
  2  1.95250         osd.2      up  1.00000          1.00000
-3  8.12239     host OSD2
  3  1.95250         osd.3    down        0          1.00000
  4  1.95250         osd.4    down        0          1.00000
  5  1.95250         osd.5    down        0          1.00000
  8  1.95250         osd.8    down        0          1.00000
  9  0.31239         osd.9      up  1.00000          1.00000

This looked alot better before that last extra power outage... =( Can't mount it anymore!
# ceph health
HEALTH_ERR 22 pgs are stuck inactive for more than 300 seconds; 44 pgs backfill_toofull; 80 pgs backfill_wait; 122 pgs degraded; 6 pgs down; 8 pgs inconsistent; 6 pgs peering; 2 pgs recovering; 18 pgs recovery_wait; 16 pgs stale; 122 pgs stuck degraded; 6 pgs stuck inactive; 16 pgs stuck stale; 159 pgs stuck unclean; 102 pgs stuck undersized; 102 pgs undersized; 1 requests are blocked > 32 sec; recovery 1803466/4503980 objects degraded (40.042%); recovery 692976/4503980 objects misplaced (15.386%); recovery 147/2251990 unfound (0.007%); 1 near full osd(s); 54 scrub errors; mds cluster is degraded; no legacy OSD present but 'sortbitwise' flag is not set

Each of osds are showing different failure signature.

I've uploaded osd log with debug osd = 20, debug filestore = 20, and debug ms = 20. You can find it in below links. Let me know if there is preferred way to share this!
https://drive.google.com/open?id=0By7YztAJNGUWQXItNzVMR281Snc (ceph-osd.3.log)
https://drive.google.com/open?id=0By7YztAJNGUWYmJBb3RvLVdSQWc (ceph-osd.4.log)
https://drive.google.com/open?id=0By7YztAJNGUWaXhRMlFOajN6M1k (ceph-osd.5.log)
https://drive.google.com/open?id=0By7YztAJNGUWdm9BWFM5a3ExOFE (ceph-osd.8.log)

So how does this look? Can this be fixed? =) If so please let me know. I used to take backups but since it grew so big, I wasn't able to do so anymore... and would like to get most of these back if I can. Please let me know if you need more info!

Thank you!

Regards,
Hong



with only 2 osd host. how are you doing replication ? i assume you use size=2, and that is somewhat ok, if you have min_size=2, but if you have min_size=1 it can quickly become a big problem of lost objects.

with size=2, min_size=2 your data should be on 2 drives safely(if you can get one of them running again), but your cluster will block when there is an issue.

if at all possible i would add a third osd node in your cluster. so your OK PG's can replicate to that and you can work on the down osd's without fear of loosing additional working osd's

Also some of your logs contains lines like...

failed to bind the UNIX domain socket to '/var/run/ceph/ceph-osd.3.asok': (17) File exists

filestore(/var/lib/ceph/osd/ceph-3) lock_fsid failed to lock /var/lib/ceph/osd/ceph-3/fsid, is another ceph-osd still running? (11) Resource temporarily unavailable

7faf16e23800 -1 osd.3 0 OSD::pre_init: object store '/var/lib/ceph/osd/ceph-3' is currently in use. (Is ceph-osd already running?)

7faf16e23800 -1  ** ERROR: osd pre_init failed: (16) Device or resource busy



This can indicate that you have a dead osd3 process keeping the resources open, and preventing a new osd from starting.

check with ps aux if you can see any ceph processes. If you do find somthging relating to your down osds's you should try stopping it normally, and if that fails. killing it manually. before trying to restart the osd.

also check dmesg if you have messages relating to faulty hardware or OOM killer there. i have had experiences with the OOM killer where the osd node became unreliable until i rebooted the machine.


kind regards, and good luck
Ronny Aasen

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to