-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

This is great work. Thanks for sharing. I'm looking forward to the
resolution of the scale out issues!
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1


On Wed, Sep 9, 2015 at 6:33 PM, Somnath Roy  wrote:
> Hi,
> Here is the updated presentation we discussed in the performance meeting 
> today with performance data incorporated for the scenario where both 
> journal/data on the same SSD.
>
> https://docs.google.com/presentation/d/15-Uqk0b4s1fVV1cG1G6Kba9xafcnIoLvfq8LUY7KBL0/edit#slide=id.p4
>
>
> Here is the summary:
> --------------------------
>
> For 4K :
> –       ~2.5X IOPs improvement, ~2.68X latency improvement with NVRAM as 
> journal
> –       ~2.2X IOPS improvement, ~2.3X latency improvement  without NVRAM as 
> journal
> –       More stable performance with NVRAM as journal
> For 16K:
> –       ~2.3X IOPS improvement, ~2.47X latency improvement with NVRAM as 
> journal
> –       ~1.9X IOPS improvement, ~2X latency improvement  without NVRAM as 
> journal
> –       More stable performance with NVRAM as journal
> For 64K:
> –       ~2X BW improvement, ~2X latency improvement with NVRAM as journal
> –       ~1.5X BW improvement, ~1.5X latency improvement  without NVRAM as 
> journal
> –       More stable performance with NVRAM as journal
> For 4M:
> –       ~1.5X BW improvement, ~1.6X latency improvement with NVRAM as journal
> –       ~1.1X BW improvement, ~1.2X latency improvement  without NVRAM as 
> journal
>
> For Mixed workload:
> ------------------------
> QD = 8 :
>         ~1.6X IOPs improvement, ~1.6X latency improvement with NVRAM as 
> journal
>         ~1.5X IOPs improvement, ~1.5X latency improvement without NVRAM as 
> journal
>         More stable performance with NVRAM as journal
>
> QD = 80 :
>         ~2.7X IOPs improvement, ~2.8X latency improvement with NVRAM as 
> journal
>         ~2.4X IOPs improvement, ~2.5X latency improvement without NVRAM as 
> journal
>         More stable performance with NVRAM as journal
>
> Will try to send out a pull request by next week.
>
> Thanks & Regards
> Somnath
>
> -----Original Message-----
> From: Somnath Roy
> Sent: Thursday, September 03, 2015 1:06 PM
> To: 'Blinick, Stephen L'
> Cc: ceph-devel
> Subject: RE: Ceph Write Path Improvement
>
> Stephen,
> It's a 1 RBD volume (preconditioned) of 2 TB size from one physical client 
> box.
> fio-rbd script I am running with 10 jobs and each with 64 QD.
> For mixed workload it is with QD = 8 and num_job= 1 and 10.
>
>
> Thanks & Regards
> Somnath
>
> -----Original Message-----
> From: Blinick, Stephen L [mailto:[email protected]]
> Sent: Thursday, September 03, 2015 1:02 PM
> To: Somnath Roy
> Cc: ceph-devel
> Subject: RE: Ceph Write Path Improvement
>
> Somnath -- thanks for publishing all the data, will be great to look at it 
> offline.  I didn't find this info:  How many RBD volumes, and what size, did 
> you use for your mixed tests?  Was it just one RBD w/ num_jobs=1 & 10?  Also 
> how many client systems were necessary to drive the workload on the 4 storage 
> nodes?
>
> I saw the same behavior quite a while back when playing with ramdisk 
> journal... Not a lot of improvement.
>
> Thanks,
>
> Stephen
>
> -----Original Message-----
> From: [email protected] 
> [mailto:[email protected]] On Behalf Of Somnath Roy
> Sent: Thursday, September 3, 2015 10:42 AM
> To: Mark Nelson; Robert LeBlanc
> Cc: ceph-devel
> Subject: RE: Ceph Write Path Improvement
>
> Yes, As Mark said I will collect all the data and hopefully I can present in 
> the next performance meeting.
> BTW, I have tested with Hammer code base + NvRAM journal initially, but that 
> performance is very spiky with ~10% performance gain (at max). I thought 
> there is no point of collecting more data with that config.
> That's why I have introduced a new throttling scheme that should benefit in 
> all the scenarios.
>
> Thanks & Regards
> Somnath
>
> -----Original Message-----
> From: Mark Nelson [mailto:[email protected]]
> Sent: Thursday, September 03, 2015 9:42 AM
> To: Robert LeBlanc; Somnath Roy
> Cc: ceph-devel
> Subject: Re: Ceph Write Path Improvement
>
> On 09/03/2015 11:23 AM, Robert LeBlanc wrote:
>> -----BEGIN PGP SIGNED MESSAGE-----
>> Hash: SHA256
>>
>> Somnath,
>>
>> I'm having a hard time with your slide deck. Am I understanding
>> correctly that the default Hammer install was performed on SSDs with
>> co-located journals, but the optimized code was performed on the same
>> SSDs but the journal was in NVRAM? If so I'm having a hard time
>> understanding how these tests can be comparable. I really like the
>> performance gains you are seeing, but I'm trying to understand how
>> much the optimized code alone helps performance.
>
> Hi Robert,
>
> We talked about this a bit at the weekly performance meeting.  I think 
> Somnath just hasn't gotten a chance to do those tests yet and is planning on 
> doing them in the coming weeks.  I believe he started out with hammer on the 
> SSDs and then tried to figure out how to tweak things to make the NVRAM 
> configuration perform better.  Now he has to go back and retest the original 
> configuration but with the new code.
>
> Mark
>
> ________________________________
>
> PLEASE NOTE: The information contained in this electronic mail message is 
> intended only for the use of the designated recipient(s) named above. If the 
> reader of this message is not the intended recipient, you are hereby notified 
> that you have received this message in error and that any review, 
> dissemination, distribution, or copying of this message is strictly 
> prohibited. If you have received this communication in error, please notify 
> the sender by telephone or e-mail (as shown above) immediately and destroy 
> any and all copies of this message in your possession (whether hard copies or 
> electronically stored copies).
>
>    칻  & ~ &    +-  ݶ   w  ˛   m    ^  b  ^n r   z    h    &    G   h  ( 階 ݢj" 
>     m     z ޖ   f   h   ~ m

-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.0.2
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJV8bvwCRDmVDuy+mK58QAAii0QAJRExHq2QlmLwI7rOD8K
JeXJ+rWJlP6CluRmuS3nN2UwT+hwxzzokyA4vzqhe0lvQkiiVPKXnuSp3hlo
U6uPjSigR0XwXdXq8nlIr51lXByB7/dbWUu62HORgi350nY1I2NzLDLUo8Hx
qxBrC4ZkH/h2+pfU2ann5nOBqe/0EM0NlbI0VQUvCo3M5il5GbqVWq/ZHDKj
LnrQtiz/QEWhq0xnJWdYlhFzkpGGNszdoHCzmSL/oGDbEpzDp6RxjF4YB1e8
Cf1xTDwmdZf3SJKKVsPbpmWG8/F3O6PKf4sbbWdo20DBYmrBbnpFWtJEr6q0
TKr2bXycSza7TNhoBNR+TGVyDuHDxnpMGneIN/NtVffwHn5dASPvveC2Y2OE
fRWccfDeaEXESfrwzMPWFk1n+ty6hpLvClyQaXenb2POmSSJ3VN9kxye8x0U
/7DiwJNWj1YaIgWGlpbOmQSWX2/PrI++MNRIB7ck0O1LG7ERhPUwdgYRahZ7
ImN8JfvOdUb7b160hdRAKpepNElfAeIkJKRlC6+6mZsOqw53Oa0MZtYMSzwU
3VbwBeJWqUS9itGCkqITNhtikdfKdWzIcwAGRPDFPqLqXWBs1twstN/ScPBa
fWDSbSjoi09LTAIsXrnLta44QZwuQDhEWUEhUcidtIbaD1t4kbCeGC+TKuA8
VluX
=/0NL
-----END PGP SIGNATURE-----
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to