>>> This would be a good thing to bring up in the meeting on Wednesday.
>yes !
>
Yes, we can discuss details on Wed's call.
>
>>>I wonder how much effect flow-control and header/data crc had.
>yes. I known that sommath also disable crc for his bench
>
I disabled ceph's header/data crc for both simplemessenger & xio but
didn't run with header/data crc enable to see the differences.
>
>>>Were the simplemessenger tests on IPoIB or native?
>
>I think it's native, as the Vu Pham benchmark was done on mellanox
>sx1012 (ethernet).
>And xio messenger was on Roce (rdma over ethernet)
>
Yes, it's native for simplemessenger and RoCE for xio messenger
>
>>>How big was the RBD volume that was created (could some data be
>>>locally cached)? Did network data transfer statistics match the
>>>benchmark result numbers?
>
Single OSD on 4GB ramdisk, journal size is 256MB.
RBD volume is only 128MB; however, I ran fio_rbd client with direct=1 to
bypass local buffer cache
Yes, the network data xfer statistics match the benchmark result
numbers.
I used "dstat -N <ethX>" to monitor the network data statistics
I also turned all cores @ full speed and applied one parameter tuning
for Mellanox ConnectX-3 HCA mlx4_core driver
(options mlx4_core log_num_mgm_entry_size=-7)
$ cat /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_max_freq
2601000
$ for c in ./cpu[0-55]*; do echo 2601000 >
${c}/cpufreq/scaling_min_freq; done
>
>
>
>I @cc Vu pham to this mail maybe it'll be able to give us answer.
>
>
>Note that I'll have same mellanox switches (sx1012) for my production
>cluster in some weeks,
>so I'll be able to reproduce the bench. (with 2x10 cores 3,1ghz nodes
>and clients).
>
>
>
>
>
>----- Mail original -----
>De: "Mark Nelson" <[email protected]>
>À: "aderumier" <[email protected]>
>Cc: "ceph-devel" <[email protected]>, "ceph-users"
><[email protected]>
>Envoyé: Lundi 2 Mars 2015 15:39:24
>Objet: Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results
>
>Hi Alex,
>
>I see I even responded in the same thread! This would be a good thing
>to bring up in the meeting on Wednesday. Those are far faster single
>OSD results than I've been able to muster with simplemessenger. I
>wonder how much effect flow-control and header/data crc had. He did
>have quite a bit more CPU (Intel specs say 14 cores @ 2.6GHz, 28 if you
>count hyperthreading). Depending on whether there were 1 or 2 CPUs in
>that node, that might be around 3x the CPU power I have here.
>
>Some other thoughts: Were the simplemessenger tests on IPoIB or native?
>How big was the RBD volume that was created (could some data be
>locally cached)? Did network data transfer statistics match the
>benchmark result numbers?
>
>I also did some tests on fdcache, though just glancing at the results
>it
>doesn't look like tweaking those parameters had much effect.
>
>Mark
>
>On 03/01/2015 08:38 AM, Alexandre DERUMIER wrote:
>> Hi Mark,
>>
>> I found an previous bench from Vu Pham (it's was about
>>simplemessenger vs xiomessenger)
>>
>> http://www.spinics.net/lists/ceph-devel/msg22414.html
>>
>> and with 1 osd, he was able to reach 105k iops with simple messenger
>>
>> . ~105k iops (4K random read, 20 cores used, numjobs=8, iopdepth=32)
>>
>> this was with more powerfull nodes, but the difference seem to be
>>quite huge
>>
>>
>>
>> ----- Mail original -----
>> De: "aderumier" <[email protected]>
>> À: "Mark Nelson" <[email protected]>
>> Cc: "ceph-devel" <[email protected]>, "ceph-users"
>><[email protected]>
>> Envoyé: Vendredi 27 Février 2015 07:10:42
>> Objet: Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results
>>
>> Thanks Mark for the results,
>> default values seem to be quite resonable indeed.
>>
>>
>> I also wonder is cpu frequency can have an impact on latency or not.
>> I'm going to benchmark on dual xeon 10-cores 3,1ghz nodes in coming
>>weeks,
>> I'll try replay your benchmark to compare
>>
>>
>>
>> ----- Mail original -----
>> De: "Mark Nelson" <[email protected]>
>> À: "ceph-devel" <[email protected]>, "ceph-users"
>><[email protected]>
>> Envoyé: Jeudi 26 Février 2015 05:44:15
>> Objet: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results
>>
>> Hi Everyone,
>>
>> In the Ceph Dumpling/Firefly/Hammer SSD/Memstore performance
>>comparison
>> thread, Alexandre DERUMIER wondered if changing the default shard and
>> threads per shard OSD settings might have a positive effect on
>> performance in our tests. I went back and used one of the PCIe SSDs
>> from our previous tests to experiment with a recent master pull. I
>> wanted to know how performance was affected by changing these
>>parameters
>> and also to validate that the default settings still appear to be
>>correct.
>>
>> I plan to conduct more tests (potentially across multiple SATA SSDs
>>in
>> the same box), but these initial results seem to show that the
>>default
>> settings that were chosen are quite reasonable.
>>
>> Mark
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>in
>> the body of a message to [email protected]
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>N�����r��y����b�X��ǧv�^�){.n�+���z�]z���{ay�ʇڙ�,j��f���h���z��w���
>>���j:+v���w�j�m��������zZ+�����ݢj"��!�i