Somnath,
on the small workload performance, 107k is higher than the theoretical IOPS
of 520, any idea why?



>>Single client is ~14K iops, but scaling as number of clients increases.
10 clients *~107K* iops. ~25 cpu cores are used.


2014-09-01 11:52 GMT+08:00 Jian Zhang <amberzhan...@gmail.com>:

> Somnath,
> on the small workload performance,
>
>
>
> 2014-08-29 14:37 GMT+08:00 Somnath Roy <somnath....@sandisk.com>:
>
>   Thanks Haomai !
>>
>> Here is some of the data from my setup.
>>
>>
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>
>> Set up:
>>
>> --------
>>
>>
>>
>> *32 core* cpu with HT enabled, 128 GB RAM, one SSD (both journal and
>> data) -> *one OSD*. 5 client m/c with 12 core cpu and each running two
>> instances of ceph_smalliobench (10 clients total). Network is 10GbE.
>>
>>
>>
>> Workload:
>>
>> -------------
>>
>>
>>
>> Small workload – 20K objects with 4K size and io_size is also *4K RR*.
>> The intent is to serve the ios from memory so that it can uncover the
>> performance problems within single OSD.
>>
>>
>>
>> Results from Firefly:
>>
>> --------------------------
>>
>>
>>
>> Single client throughput is ~14K iops, but as the number of client
>> increases the aggregated throughput is not increasing. 10 clients *~15K*
>> iops. ~9-10 cpu cores are used.
>>
>>
>>
>> Result with latest master:
>>
>> ------------------------------
>>
>>
>>
>> Single client is ~14K iops, but scaling as number of clients increases.
>> 10 clients *~107K* iops. ~25 cpu cores are used.
>>
>>
>>
>>
>> --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>
>>
>>
>>
>>
>> More realistic workload:
>>
>> -----------------------------
>>
>> Let’s see how it is performing while > 90% of the ios are served from
>> disks
>>
>>  Setup:
>>
>> -------
>>
>> 40 cpu core server as a cluster node (single node cluster) with 64 GB
>> RAM. 8 SSDs -> *8 OSDs*. One similar node for monitor and rgw. Another
>> node for client running fio/vdbench. 4 rbds are configured with ‘noshare’
>> option. 40 GbE network
>>
>>
>>
>> Workload:
>>
>> ------------
>>
>>
>>
>> 8 SSDs are populated , so, 8 * 800GB = *~6.4 TB* of data.  Io_size = *4K
>> RR*.
>>
>>
>>
>> Results from Firefly:
>>
>> ------------------------
>>
>>
>>
>> Aggregated output while 4 rbd clients stressing the cluster in parallel
>> is *~20-25K IOPS* , cpu cores used ~8-10 cores (may be less can’t
>> remember precisely)
>>
>>
>>
>> Results from latest master:
>>
>> --------------------------------
>>
>>
>>
>> Aggregated output while 4 rbd clients stressing the cluster in parallel
>> is *~120K IOPS* , cpu is 7% idle i.e  ~37-38 cpu cores.
>>
>>
>>
>> Hope this helps.
>>
>>
>>
>> Thanks & Regards
>>
>> Somnath
>>
>>
>>
>> -----Original Message-----
>> From: Haomai Wang [mailto:haomaiw...@gmail.com]
>> Sent: Thursday, August 28, 2014 8:01 PM
>> To: Somnath Roy
>> Cc: Andrey Korolyov; ceph-users@lists.ceph.com
>> Subject: Re: [ceph-users] [Single OSD performance on SSD] Can't go over
>> 3, 2K IOPS
>>
>>
>>
>> Hi Roy,
>>
>>
>>
>> I already scan your merged codes about "fdcache" and "optimizing for
>> lfn_find/lfn_open", could you give some performance improvement data about
>> it? I fully agree with your orientation, do you have any update about it?
>>
>>
>>
>> As for messenger level, I have some very early works on it(
>> https://github.com/yuyuyu101/ceph/tree/msg-event), it contains a new
>> messenger implementation which support different event mechanism.
>>
>> It looks like at least one more week to make it work.
>>
>>
>>
>> On Fri, Aug 29, 2014 at 5:48 AM, Somnath Roy <somnath....@sandisk.com>
>> wrote:
>>
>> > Yes, what I saw the messenger level bottleneck is still huge !
>>
>> > Hopefully RDMA messenger will resolve that and the performance gain
>> will be significant for Read (on SSDs). For write we need to uncover the
>> OSD bottlenecks first to take advantage of the improved upstream.
>>
>> > What I experienced that till you remove the very last bottleneck the
>> performance improvement will not be visible and that could be confusing
>> because you might think that the upstream improvement you did is not valid
>> (which is not).
>>
>> >
>>
>> > Thanks & Regards
>>
>> > Somnath
>>
>> > -----Original Message-----
>>
>> > From: Andrey Korolyov [mailto:and...@xdel.ru <and...@xdel.ru>]
>>
>> > Sent: Thursday, August 28, 2014 12:57 PM
>>
>> > To: Somnath Roy
>>
>> > Cc: David Moreau Simard; Mark Nelson; ceph-users@lists.ceph.com
>>
>> > Subject: Re: [ceph-users] [Single OSD performance on SSD] Can't go
>>
>> > over 3, 2K IOPS
>>
>> >
>>
>> > On Thu, Aug 28, 2014 at 10:48 PM, Somnath Roy <somnath....@sandisk.com>
>> wrote:
>>
>> >> Nope, this will not be back ported to Firefly I guess.
>>
>> >>
>>
>> >> Thanks & Regards
>>
>> >> Somnath
>>
>> >>
>>
>> >
>>
>> > Thanks for sharing this, the first thing in thought when I looked at
>>
>> > this thread, was your patches :)
>>
>> >
>>
>> > If Giant will incorporate them, both the RDMA support and those should
>> give a huge performance boost for RDMA-enabled Ceph backnets.
>>
>> >
>>
>> > ________________________________
>>
>> >
>>
>> > PLEASE NOTE: The information contained in this electronic mail message
>> is intended only for the use of the designated recipient(s) named above. If
>> the reader of this message is not the intended recipient, you are hereby
>> notified that you have received this message in error and that any review,
>> dissemination, distribution, or copying of this message is strictly
>> prohibited. If you have received this communication in error, please notify
>> the sender by telephone or e-mail (as shown above) immediately and destroy
>> any and all copies of this message in your possession (whether hard copies
>> or electronically stored copies).
>>
>> >
>>
>> > _______________________________________________
>>
>> > ceph-users mailing list
>>
>> > ceph-users@lists.ceph.com
>>
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>>
>>
>>
>>
>> --
>>
>> Best Regards,
>>
>>
>>
>> Wheat
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to