On Thu, Apr 17, 2014 at 6:21 AM, Qing Zheng <[email protected]> wrote:
> I seems that with kernel 3.14 and latest source code from github, we still
> run
> into troubles when testing multi-mds and directory sharding.
>
what's problem you encountered?




> Are there any limits either in the max number of active metadata servers
> that we could possibly run
> or in the number of directory entries that we could set for Ceph to trigger
> a directory split?
>
> Is it okay to run 128 or more active metadata servers, for example?
> Is it okay to let Ceph split directories once a directory has accumulated
> 200 entries?
>
> Cheers,
>
> -- Qing Zheng
>
> -----Original Message-----
> From: Yan, Zheng [mailto:[email protected]]
> Sent: Sunday, April 13, 2014 6:43 PM
> To: Qing Zheng
> Cc: [email protected]
> Subject: Re: [ceph-users] multi-mds and directory sharding
>
> On Mon, Apr 14, 2014 at 2:54 AM, Qing Zheng <[email protected]> wrote:
>> Hi -
>>
>> We are currently evaluating CephFS's metadata scalability and
>> performance. One important feature of CephFS is its support for
>> running multiple "active" mds instances and partitioning huge
>> directories into small shards.
>>
>> We use mdtest to simulate workloads where multiple parallel client
>> processes will keep inserting empty files into several large directories.
>> We found that CephFS is only able to run for the first 5-10 mins, and
>> then stop making progress -- the clients' "creat" call no longer return.
>>
>> We were using Ceph 0.72 and Ubuntu 12.10 with kernel 3.6.6.
>> Our setup consisted of 8 osds, 3 mds, and 1 mon. All mds were active,
>> instead of standby, and they were all configured to split directories
>> once the directory size is greater than 2k. We kernel (not fuse)
>> mounted CephFS on all 8 osd nodes.
>
> 3.6 kernel is too old for cephfs. please use kernel compiled from testing
> branch https://github.com/ceph/ceph-client and the newest development
> version of Ceph. There are large number of fixes for directory fragment and
> multimds.
>
> Regards
> Yan, Zheng
>
>>
>> To test CephFS, we launched 64 client processes on 8 osd nodes (8
>> procs per osd). Each client would create 1 directory and then insert
>> 5k empty files into that directory. In total 64 directories and 320k
>> files would be created. CephFS gave an avg throughput of 300~1k for
>> the first 5 minutes, and then stopped making any progress.
>>
>> What might go wrong?
>>
>> If each client insert 200 files, instead of 5k, then CephFS could
>> finish the workload with 1.5K ops/s. If each client insert 1k files,
>> then ~500 ops/s If 2k files (the split threshold), then ~400 ops/s
>>
>> Are these numbers reasonable?
>>
>> -- Qing
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to