Cinder will periodically inspect the free space of the volume services and
use this data when determining which one to schedule to when a request is
received. In this case the cinder volume create request may error out in
scheduling. You may also see an error when instantiating a volume from an
image if it passes the prior but then becomes out of space during writing
the image to the volume.

I'm not sure if it's still the case, but in Havana (I see no reason for it
to change) the free space check in cinder didn't account for the difference
between promised space (the max of the volumes assigned) instead it would
literally look for free space in the output of `rados df`

As noted above if the cluster gets to "100%" used, bad things will happen
to your VM's. The most likely case is that they all assert read-only
filesystems. (100% is a missnomer as there is a configured max % where it
will stop accepting data writes to ensure that important object replication
/ maintenance can occur and have the cluster not fall over)

On Mon, Oct 19, 2015 at 7:51 AM LOPEZ Jean-Charles <[email protected]>
wrote:

> Hi,
>
> when an OSD gets full, any write operation to the entire cluster will be
> disabled.
>
> As a result, creating a single RBD will become impossible and all VMs that
> need to write to one of their Ceph back RBDs will suffer the same pain.
>
> Usually, this ends up as a bad sorry for the VMs.
>
> The best practice is to monitor the disk space usage for the OSDs and as a
> matter of fact RHCS 1.# includes a cep old df command to do this. You can
> also use the output of the cep old report command to grab the appropriate
> info to compute it or rely on external SNMP monitoring tools to grab the
> usage details of the particular OSD disk drives.
>
> Have a great day.
> JC
>
> > On Oct 19, 2015, at 02:32, Bharath Krishna <[email protected]>
> wrote:
> >
> > I mean cluster OSDs are physically full.
> >
> > I understand its not a pretty way to operate CEPH allowing to become
> full,
> > but I just wanted to know the boundary condition if it becomes full.
> >
> > Will cinder create volume operation creates new volume at all or error is
> > thrown at Cinder API level itself stating that no space available?
> >
> > When IO stalls, will I be able to read the data from CEPH cluster I.e can
> > I still read data from existing volumes created from CEPH cluster?
> >
> > Thanks for the quick reply.
> >
> > Regards
> > M Bharath Krishna
> >
> > On 10/19/15, 2:51 PM, "Jan Schermer" <[email protected]> wrote:
> >
> >> Do you mean when the CEPH cluster (OSDs) is physically full or when the
> >> quota is reached?
> >>
> >> If CEPH becomes full it just stalls all IO (maybe just write IO, but
> >> effectively same thing) - not pretty and you must never ever let it
> >> become full.
> >>
> >> Jan
> >>
> >>
> >>> On 19 Oct 2015, at 11:15, Bharath Krishna <[email protected]>
> >>> wrote:
> >>>
> >>> Hi
> >>>
> >>> What happens when Cinder service with CEPH backend storage cluster
> >>> capacity is FULL?
> >>>
> >>> What would be the out come of new cinder create volume request?
> >>>
> >>> Will volume be created with space not available for use or an error
> >>> thrown from Cinder API stating no space available for new volume.
> >>>
> >>> I could not try this in my environment and fill up the cluster.
> >>>
> >>> Please reply if you have ever tried and tested this.
> >>>
> >>> Thank you.
> >>>
> >>> Regards,
> >>> M Bharath Krishna
> >>> _______________________________________________
> >>> ceph-users mailing list
> >>> [email protected]
> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>
> >
> > _______________________________________________
> > ceph-users mailing list
> > [email protected]
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
-- 

--

Andrew Woodward

Mirantis

Fuel Community Ambassador

Ceph Community
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to