On Tue, Jun 30, 2015 at 11:37 AM, Yan, Zheng <[email protected]> wrote:
>
>> On Jun 30, 2015, at 15:37, Ilya Dryomov <[email protected]> wrote:
>>
>> On Tue, Jun 30, 2015 at 6:57 AM, Yan, Zheng <[email protected]> wrote:
>>> I tried 4.1 kernel and 0.94.2 ceph-fuse. their performance are about the
>>> same.
>>>
>>> fuse:
>>> Files=191, Tests=1964, 60 wallclock secs ( 0.43 usr 0.08 sys + 1.16 cusr
>>> 0.65 csys = 2.32 CPU)
>>>
>>> kernel:
>>> Files=191, Tests=2286, 61 wallclock secs ( 0.45 usr 0.08 sys + 1.21 cusr
>>> 0.72 csys = 2.46 CPU)
>>
>> On Friday, I tried stock 3.10 vs 4.1 and they were about the same as
>> well (a few tests failed in 3.10 though). However Dan is using
>> 3.10.0-229.7.2.el7.x86_64, which is 3.10 with a lot of backports, so
>> it's not quite the same. Dan, are the numbers you are seeing
>> consistent?
>>
>
> I just tried 3.10.0-229.7.2.el7 kernel. it’s a little slower than 4.1 kernel
>
> 4.1:
> Files=191, Tests=2286, 61 wallclock secs ( 0.45 usr 0.07 sys + 1.24 cusr
> 0.76 csys = 2.52 CPU)
>
> 3.10.0-229.7.2.el7:
> Files=191, Tests=1964, 75 wallclock secs ( 0.45 usr 0.09 sys + 1.73 cusr
> 5.04 csys = 7.31 CPU)
>
> Dan, did you run the test on the same client machine. I think network latency
> affects run time of this test a lots
>
All the tests run on the same client, but it seems there is some
variability in the tests. Now I get:
Linux 3.10.0-229.7.2.el7.x86_64
Files=184, Tests=1957, 91 wallclock secs ( 0.72 usr 0.19 sys + 5.68
cusr 10.09 csys = 16.68 CPU)
Linux 4.1.0-1.el7.elrepo.x86_64
Files=184, Tests=1957, 84 wallclock secs ( 0.75 usr 0.44 sys + 5.17
cusr 9.77 csys = 16.13 CPU)
ceph-fuse 0.94.2:
Files=184, Tests=1957, 78 wallclock secs ( 0.69 usr 0.17 sys + 5.08
cusr 9.93 csys = 15.87 CPU)
I don't know if it's related -- and maybe I misunderstood something
fundamental -- but we don't manage to get FUSE or the kernel client to
use the page cache:
I have fuse_use_invalidate_cb = true then used fincore to see what's cached:
# df -h .
Filesystem Size Used Avail Use% Mounted on
ceph-fuse 444T 135T 309T 31% /cephfs
# cat zero > /dev/null
# linux-fincore zero
filename
size total_pages min_cached page
cached_pages cached_size cached_perc
--------
---- ----------- ---------------
------------ ----------- -----------
zero
104,857,600 25,600 -1
0 0 0.00
---
total cached size: 0
The kernel client has the same behaviour. Is this expected?
Cheers, Dan
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com