Hi, >>Thanks for the detailed information. but I am already using fio with rbd >>engine. Almost 4 volumes can reach the peak.
What is your cpu usage of fio-rbd ? Myself I'm cpu bound on 8cores with around 40000iops read 4K. ----- Mail original ----- De: "Mark Wu" <[email protected]> À: "Daniel Schwager" <[email protected]> Cc: [email protected] Envoyé: Jeudi 16 Octobre 2014 19:19:17 Objet: Re: [ceph-users] Performance doesn't scale well on a full ssd cluster. Thanks for the detailed information. but I am already using fio with rbd engine. Almost 4 volumes can reach the peak. 2014 年 10 月 17 日 上午 1:03于 [email protected] 写道: Thanks for the detailed information. but I am already using fio with rbd engine. Almost 4 volumes can reach the peak. 2014 年 10 月 17 日 上午 12:55于 "Daniel Schwager" < [email protected] >写道: <blockquote> Hi Mark, maybe you will check rbd-enabled fio http://telekomcloud.github.io/ceph/2014/02/26/ceph-performance-analysis_fio_rbd.html yum install ceph-devel git clone git:// git.kernel.dk/fio.git cd fio ; ./configure ; make -j5 ; make install Setup the number of jobs (==clients) inside fio config to numjobs=8 for simulating multiple clients. regards Danny my test.fio: [global] #logging #write_iops_log=write_iops_log #write_bw_log=write_bw_log #write_lat_log=write_lat_log ioengine=rbd clientname=admin pool=rbd rbdname=myimage invalidate=0 # mandatory rw=randwrite bs=1m runtime=120 iodepth=8 numjobs=8 time_based #direct=0 [seq-write] stonewall rw=write #[seq-read] #stonewall #rw=read </blockquote> _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
