Yes, I read that and I investigated all those areas, on my Dumpling the gains 
are pretty low.
If I remember correctly Hammer didn't improve synchronous (journal) writes at 
all. Or at least it didn't when I read that...?
So is it actually that much faster? Did something change in Hammer in recent 
releases?
3x speedup is of course good, but it still needs to get even faster. I'd 
consider 0.5ms in all situations a bottom line goal to be able to run 
moderately loaded databases, but to compete with any SAN or even DAS storage it 
needs to get _much_ faster than that, somehow...
Otherwise we're still just wasting SSDs in here.

Jan

> On 10 Sep 2015, at 16:45, Haomai Wang <[email protected]> wrote:
> 
> 
> 
> On Thu, Sep 10, 2015 at 10:41 PM, Jan Schermer <[email protected] 
> <mailto:[email protected]>> wrote:
> What did you tune? Did you have to make a human sacrifice? :) Which release?
> The last proper benchmark numbers I saw were from hammer and the latencies 
> were basically still the same, about 2ms for write.
> 
> No sacrifice, actually I dive into all-ssd ceph since 2013, I can see the 
> improvement from Dumpling to Hammer. 
> 
> You can find my related thread in 2014. Mainly about ensure fd hit, cpu 
> powersave disable, memory management
>  
> 
> Jan
> 
> 
>> On 10 Sep 2015, at 16:38, Haomai Wang <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>> 
>> 
>> On Thu, Sep 10, 2015 at 10:36 PM, Jan Schermer <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>>> On 10 Sep 2015, at 16:26, Haomai Wang <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> 
>>> Actually we can reach 700us per 4k write IO for single io depth(2 copy, 
>>> E52650, 10Gib, intel s3700). So I think 400 read iops shouldn't be a 
>>> unbridgeable problem.
>>> 
>> 
>> Flushed to disk?
>> 
>> of course
>>  
>> 
>> 
>>> CPU is critical for ssd backend, so what's your cpu model?
>>> 
>>> On Thu, Sep 10, 2015 at 9:48 PM, Jan Schermer <[email protected] 
>>> <mailto:[email protected]>> wrote:
>>> It's certainly not a problem with DRBD (yeah, it's something completely 
>>> different but it's used for all kinds of workloads including things like 
>>> replicated tablespaces for databases).
>>> It won't be a problem with VSAN (again, a bit different, but most people 
>>> just want something like that)
>>> It surely won't be a problem with e.g. ScaleIO which should be comparable 
>>> to Ceph.
>>> 
>>> Latency on the network can be very low (0.05ms on my 10GbE). Latency on 
>>> good SSDs is  2 orders of magnitute lower (as low as 0.00005 ms). Linux is 
>>> pretty good nowadays at waking up threads and pushing the work. Multiply 
>>> those numbers by whatever factor and it's still just a fraction of the 
>>> 0.5ms needed.
>>> The problem is quite frankly slow OSD code and the only solution now is to 
>>> keep the data closer to the VM.
>>> 
>>> Jan
>>> 
>>> > On 10 Sep 2015, at 15:38, Gregory Farnum <[email protected] 
>>> > <mailto:[email protected]>> wrote:
>>> >
>>> > On Thu, Sep 10, 2015 at 2:34 PM, Stefan Priebe - Profihost AG
>>> > <[email protected] <mailto:[email protected]>> wrote:
>>> >> Hi,
>>> >>
>>> >> while we're happy running ceph firefly in production and also reach
>>> >> enough 4k read iop/s for multithreaded apps (around 23 000) with qemu 
>>> >> 2.2.1.
>>> >>
>>> >> We've now a customer having a single threaded application needing around
>>> >> 2000 iop/s but we don't go above 600 iop/s in this case.
>>> >>
>>> >> Any tuning hints for this case?
>>> >
>>> > If the application really wants 2000 sync IOPS to disk without any
>>> > parallelism, I don't think any network storage system is likely to
>>> > satisfy him — that's only half a millisecond per IO. 600 IOPS is about
>>> > the limit of what the OSD can do right now (in terms of per-op
>>> > speeds), and although there is some work being done to improve that
>>> > it's not going to be in a released codebase for a while.
>>> >
>>> > Or perhaps I misunderstood the question?
>>> > _______________________________________________
>>> > ceph-users mailing list
>>> > [email protected] <mailto:[email protected]>
>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>>> > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>> 
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected] <mailto:[email protected]>
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>>> 
>>> 
>>> 
>>> -- 
>>> Best Regards,
>>> 
>>> Wheat
>>> 
>> 
>> 
>> 
>> 
>> -- 
>> Best Regards,
>> 
>> Wheat
>> 
> 
> 
> 
> 
> -- 
> Best Regards,
> 
> Wheat
> 

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to