yes, but: # rbd showmapped id pool image snap device 0 rbd nfs1 - /dev/rbd0 1 rbd nfs2 - /dev/rbd1
# df -h Filesystem Size Used Avail Use% Mounted on /dev/rbd0 8.0T 4.8T 3.3T 60% /mnt/nfsroot/rbd0 /dev/rbd1 9.8T 34M 9.8T 1% /mnt/nfsroot/rbd1 only 5T is taken up On Thu, Feb 28, 2019 at 2:26 PM Jack <[email protected]> wrote: > Are not you using 3-replicas pool ? > > (15745GB + 955GB + 1595M) * 3 ~= 51157G (there is overhead involved) > > Best regards, > > On 02/28/2019 11:09 PM, solarflow99 wrote: > > thanks, I still can't understand whats taking up all the space 27.75 > > > > On Thu, Feb 28, 2019 at 7:18 AM Mohamad Gebai <[email protected]> wrote: > > > >> On 2/27/19 4:57 PM, Marc Roos wrote: > >>> They are 'thin provisioned' meaning if you create a 10GB rbd, it does > >>> not use 10GB at the start. (afaik) > >> > >> You can use 'rbd -p rbd du' to see how much of these devices is > >> provisioned and see if it's coherent. > >> > >> Mohamad > >> > >>> > >>> > >>> -----Original Message----- > >>> From: solarflow99 [mailto:[email protected]] > >>> Sent: 27 February 2019 22:55 > >>> To: Ceph Users > >>> Subject: [ceph-users] rbd space usage > >>> > >>> using ceph df it looks as if RBD images can use the total free space > >>> available of the pool it belongs to, 8.54% yet I know they are created > >>> with a --size parameter and thats what determines the actual space. I > >>> can't understand the difference i'm seeing, only 5T is being used but > >>> ceph df shows 51T: > >>> > >>> > >>> /dev/rbd0 8.0T 4.8T 3.3T 60% /mnt/nfsroot/rbd0 > >>> /dev/rbd1 9.8T 34M 9.8T 1% /mnt/nfsroot/rbd1 > >>> > >>> > >>> > >>> # ceph df > >>> GLOBAL: > >>> SIZE AVAIL RAW USED %RAW USED > >>> 180T 130T 51157G 27.75 > >>> POOLS: > >>> NAME ID USED %USED MAX AVAIL > >>> OBJECTS > >>> rbd 0 15745G 8.54 39999G > >>> 4043495 > >>> cephfs_data 1 0 0 39999G > >>> 0 > >>> cephfs_metadata 2 1962 0 39999G > >>> 20 > >>> spider_stage 9 1595M 0 39999G 47835 > >>> spider 10 955G 0.52 39999G > >>> 42541237 > >>> > >>> > >>> > >>> > >>> _______________________________________________ > >>> ceph-users mailing list > >>> [email protected] > >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > >> > > > > > > > > _______________________________________________ > > ceph-users mailing list > > [email protected] > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
