mon_osd_min_down_reporters by default set to 2

I guess you’ll have to set it to 1 in your case

JC

> On Sep 29, 2016, at 08:16, Gregory Farnum <[email protected]> wrote:
> 
> I think the problem is that Ceph requires a certain number of OSDs or a 
> certain number of reports of failure before it marks an OSD down. These 
> thresholds are not tuned for a 2-OSD cluster; you probably want to set them 
> to 1.
> Also keep in mind that the OSDs provide a grace period of 20-30 seconds 
> before they'll report somebody down; this helps prevent spurious recovery but 
> means you will get paused IO on an unclean shutdown.
> 
> I can't recall the exact config options off-hand, but it's something like 
> "mon osd min down reports". Search the docs for that. :)
> -Greg
> 
> On Thursday, September 29, 2016, Peter Maloney 
> <[email protected] 
> <mailto:[email protected]>> wrote:
> On 09/29/16 14:07, Ranjan Ghosh wrote:
> > Wow. Amazing. Thanks a lot!!! This works. 2 (hopefully) last questions
> > on this issue:
> >
> > 1) When the first node is coming back up, I can just call "ceph osd up
> > 0" and Ceph will start auto-repairing everything everything, right?
> > That is, if there are e.g. new files that were created during the time
> > the first node was down, they will (sooner or later) get replicated
> > there?
> Nope, there is no "ceph osd up <id>"; you just start the osd, and it
> already gets recognized as up. (if you don't like this, you set it out,
> not just down; and there is a "ceph osd in <id>" to undo that.)
> >
> > 2) If I don't call "osd down" manually (perhaps at the weekend when
> > I'm not at the office) when a node dies - did I understand correctly
> > that the "hanging" I experienced is temporary and that after a few
> > minutes (don't want to try out now) the node should also go down
> > automatically?
> I believe so, yes.
> 
> Also, FYI, RBD images don't seem to have this issue, and work right away
> on a 3 osd cluster. Maybe cephfs would also work better with a 3rd osd,
> even an empty one (weight=0). (and I had an unresolved issue testing the
> same with cephfs on my virtual test cluster)
> >
> > BR,
> > Ranjan
> >
> >
> > Am 29.09.2016 um 13:00 schrieb Peter Maloney:
> >>
> >> And also you could try:
> >>      ceph osd down <osd id>
> >
> 
> _______________________________________________
> ceph-users mailing list
> [email protected] <javascript:;>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to