This seems like a good feature to have. I've created
http://tracker.ceph.com/issues/5642

N


On Tue, Jul 16, 2013 at 8:05 AM, Greg Chavez <[email protected]> wrote:

> This is interesting.  So there are no built-in ceph commands that can
> calculate your usable space?  It just so happened that I was going to
> try and figure that out today (new Openstack block cluster, 20TB total
> capacity) by skimming through the documentation.  I figured that there
> had to be a command that would do this.  Blast and gadzooks.
>
> On Tue, Jul 16, 2013 at 10:37 AM, Ta Ba Tuan <[email protected]> wrote:
> >
> > Thank Sage,
> >
> > tuantaba
> >
> >
> > On 07/16/2013 09:24 PM, Sage Weil wrote:
> >>
> >> On Tue, 16 Jul 2013, Ta Ba Tuan wrote:
> >>>
> >>> Thanks  Sage,
> >>> I wories about returned capacity when mounting CephFS.
> >>> but when disk is full, capacity will showed 50% or 100% Used?
> >>
> >> 100%.
> >>
> >> sage
> >>
> >>>
> >>> On 07/16/2013 11:01 AM, Sage Weil wrote:
> >>>>
> >>>> On Tue, 16 Jul 2013, Ta Ba Tuan wrote:
> >>>>>
> >>>>> Hi everyone.
> >>>>>
> >>>>> I have 83 osds, and every osds have same 2TB, (Capacity sumary is
> >>>>> 166TB)
> >>>>> I'm using replicate 3 for pools ('data','metadata').
> >>>>>
> >>>>> But when mounting Ceph filesystem from somewhere (using: mount -t
> ceph
> >>>>> Monitor_IP:/ /ceph -o name=admin,secret=xxxxxxxxxx")
> >>>>> then capacity sumary is showed "160TB"?, I used replicate 3 and I
> think
> >>>>> that
> >>>>> it must return 160TB/3=50TB?
> >>>>>
> >>>>> Filesystem                Size  Used Avail Use% Mounted on
> >>>>> 192.168.32.90:/    160T  500G  156T   1%  /tmp/ceph_mount
> >>>>>
> >>>>> Please, explain this  help me?
> >>>>
> >>>> statfs/df show the raw capacity of the cluster, not the usable
> capacity.
> >>>> How much data you can store is a (potentially) complex function of
> your
> >>>> CRUSH rules and replication layout.  If you store 1TB, you'll notice
> the
> >>>> available space will go down by about 2TB (if you're using the default
> >>>> 2x).
> >>>>
> >>>> sage
> >>>
> >>>
> >
> > _______________________________________________
> > ceph-users mailing list
> > [email protected]
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> --
> \*..+.-
> --Greg Chavez
> +//..;};
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to