Mounting and unmount doesn't change anyting.
The used space reported by df command is nearly the same as the values
returned by ceph -s command.
Example 1, df output:
ceph-fuse 334T 134T 200T 41% /cephfs
Example 2, ceph -s output:
health HEALTH_WARN
mds0: Many clients (22) failing to respond to cache pressure
noscrub,nodeep-scrub,sortbitwise flag(s) set
monmap e1: 5 mons at
{r730-12=10.103.213.12:6789/0,r730-4=10.103.213.4:6789/0,r730-5=
10.103.213.5:6789/0,r730-8=10.103.213.8:6789/0,r730-9=10.103.213.9:6789/0}
election epoch 132, quorum 0,1,2,3,4
r730-4,r730-5,r730-8,r730-9,r730-12
mdsmap e14637: 1/1/1 up {0=ceph2-mds-2=up:active}
osdmap e6549: 68 osds: 68 up, 68 in
flags noscrub,nodeep-scrub,sortbitwise
pgmap v4394151: 896 pgs, 3 pools, 54569 GB data, 56582 kobjects
133 TB used, 199 TB / 333 TB avail
896 active+clean
client io 47395 B/s rd, 1979 kB/s wr, 388 op/s
________________________________________
From: John Spray <[email protected]>
Sent: Tuesday, April 19, 2016 22:04
To: Simion Rad
Cc: [email protected]
Subject: Re: [ceph-users] cephfs does not seem to properly free up space
On Tue, Apr 19, 2016 at 2:40 PM, Simion Rad <[email protected]> wrote:
> Hello,
>
>
> At my workplace we have a production cephfs cluster (334 TB on 60 OSDs)
> which was recently upgraded from Infernalis 9.2.0 to Infernalis 9.2.1 on
> Ubuntu 14.04.3 (linux 3.19.0-33).
>
> It seems that cephfs still doesn't free up space at all or at least that's
> what df command tells us.
Hmm, historically there were bugs with the purging code, but I thought
we fixed them before Infernalis.
Does the space get freed after you unmount the client? Some issues
have involved clients holding onto references to unlinked inodes.
John
>
> Is there a better way of getting a df-like output with other command for
> cephfs ?
>
>
> Thank you,
>
> Marius Rad
>
> SysAdmin
>
> www.propertyshark.com
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com