On Jun 11, 2011, at 4:24 PM, Wilfrid Allembrand wrote:

> Hi,
> 
> So I have my small test cluster with 2 OSD nodes. In each OSD node,
> there is one disk (5 GB) and one cosd.
> When I mount the FS on a client, I see it's 10 GB in total. Fine.
> I copied a small set of multimedia files (135 MB) and that's all. So
> why do I see the space used is getting higher and higher ?
> 
> root@client:/mnt/osd$ df -h .
> Filesystem            Size  Used Avail Use% Mounted on
> 10.1.56.232:6789:/     10G  5.2G  3.9G  57% /mnt/osd
> 
> and like 2 hours later :
> 
> root@client:/mnt/osd$ df -h .
> Filesystem            Size  Used Avail Use% Mounted on
> 10.1.56.232:6789:/     10G  6.8G  2.3G  76% /mnt/osd
> 
> 
> I'm using btrfs to store the files and I made no creation/modification
> of files; i just run some bench with "ceph osd tell [12] bench"
> Does the bench create files and do note delete them ?
The OSD bench that you're invoking definitely does delete files. You shouldn't 
be seeing large growth like that with your ceph.conf, though.

Are you sure there's nothing else accessing your mount and doing writes to it? 
What I'd expect to see from 135MB of writes is ~570MB used (135MB written on 
each OSD, for the replication, doubled because you've got a 512MB journal 
recording the data too; and then plus a little bit extra for the MDS journaling 
and data).

You're not using the rados bench, are you? That one doesn't clean up after 
itself.


> 
> Extract of ceph -w :
> root@test2:/data/mon1# ceph -w
> 2011-06-11 19:16:09.186416 7fbefd01e700 -- :/14931 >>
> 10.1.56.233:6789/0 pipe(0x1e58540 sd=4 pgs=0 cs=0 l=0).fault first
> fault
> 2011-06-11 19:16:12.191439    pg v1990: 594 pgs: 594 active+clean;
> 2066 MB data, 6906 MB used, 2286 MB / 10240 MB avail
> 2011-06-11 19:16:12.193479   mds e48: 1/1/1 up {0=test4=up:active}, 1 
> up:standby
> 2011-06-11 19:16:12.193551   osd e44: 2 osds: 2 up, 2 in
> 2011-06-11 19:16:12.193658   log 2011-06-12 00:55:42.389017 osd2
> 10.1.56.237:6800/1744 307 : [INF] bench: wrote 1024 MB in blocks of
> 4096 KB in 12.438478 sec at 84300 KB/sec
> 2011-06-11 19:16:12.193757   mon e1: 3 mons at
> {0=10.1.56.231:6789/0,1=10.1.56.232:6789/0,2=10.1.56.233:6789/0}
> 
> 
> Attached is my ceph.conf
> 
> Thanks !
> Wilfrid
> <ceph.conf>

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to