On Wed, 7 Jan 2015 17:07:46 -0800 Craig Lewis wrote:

> On Mon, Dec 29, 2014 at 4:49 PM, Alexandre Oliva <ol...@gnu.org> wrote:
> 
> > However, I suspect that temporarily setting min size to a lower number
> > could be enough for the PGs to recover.  If "ceph osd pool <pool> set
> > min_size 1" doesn't get the PGs going, I suppose restarting at least
> > one of the OSDs involved in the recovery, so that they PG undergoes
> > peering again, would get you going again.
> >
> 
> It depends on how incomplete your incomplete PGs are.
> 
> min_size is defined as "Sets the minimum number of replicas required for
> I/O.".  By default, size is 3 and min_size is 2 on recent versions of
> ceph.
> 
> If the number of replicas you have drops below min_size, then Ceph will
> mark the PG as incomplete.  As long as you have one copy of the PG, you
> can recover by lowering the min_size to the number of copies you do
> have, then restoring the original value after recovery is complete.  I
> did this last week when I deleted the wrong PGs as part of a toofull
> experiment.
> 
Which of course begs the question of why not having min_size at 1
permanently, so that in the (hopefully rare) case of loosing 2 OSDs at the
same time your cluster still keeps working (as it should with a size of 3).

Christian
-- 
Christian Balzer        Network/Systems Engineer                
ch...@gol.com           Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to