Hi,
在 2013-3-25,17:30,"Wolfgang Hennerbichler"
<[email protected]> 写道:
> Hi Xiaoxi,
>
> sorry, I thought you were testing within VMs and caching turned on (I
> assumed, you didn't tell us if you really did use your benchmark within
> vms and if not, how you tested rbd outside of VMs).
Yes,I really testing within VMs
> It just triggered an alarm in me because we had also experienced issues
> with benchmarking within a VM (it didn't crash but responded extremely
> slow).
>
OK,but my VM didnt crash, it's ceph-osd daemon crashed. So is it safe for me to
say the issue I hit is a different issue?(not #3737)
> Wolfgang
xiaoxi
>
> On 03/25/2013 10:15 AM, Chen, Xiaoxi wrote:
>>
>>
>> Hi Wolfgang,
>>
>> Thanks for the reply,but why my problem is related with issue#3737? I
>> cannot find any direct link between them. I didnt turn on qemu cache and my
>> qumu/VM work fine
>>
>>
>> Xiaoxi
>>
>> 在 2013-3-25,17:07,"Wolfgang Hennerbichler"
>> <[email protected]> 写道:
>>
>>> Hi,
>>>
>>> this could be related to this issue here and has been reported multiple
>>> times:
>>>
>>> http://tracker.ceph.com/issues/3737
>>>
>>> In short: They're working on it, they know about it.
>>>
>>> Wolfgang
>>>
>>> On 03/25/2013 10:01 AM, Chen, Xiaoxi wrote:
>>>> Hi list,
>>>>
>>>> We have hit and reproduce this issue for several times, ceph
>>>> will suicide because FileStore: sync_entry timed out after a very heavy
>>>> random IO on top of the RBD.
>>>>
>>>> My test environment is:
>>>>
>>>> 4 Nodes ceph cluster with 20 HDDs for OSDs
>>>> and 4 Intel DCS3700 ssds for journal per node, that is 80 spindles in total
>>>>
>>>> 48 VMs spread across 12 Physical nodes, 48
>>>> RBD attached to the VMs 1:1 via Qemu.
>>>>
>>>> Ceph @ 0.58
>>>>
>>>> XFS were used.
>>>>
>>>> I am using Aiostress (something like FIO) to produce random
>>>> write requests on top of each RBDs.
>>>>
>>>>
>>>>
>>>> From Ceph-w , ceph reports a very high Ops (10000+ /s) , but
>>>> technically , 80 spindles can provide up to 150*80/2=6000 IOPS for 4K
>>>> random write.
>>>>
>>>> When digging into the code, I found that the OSD write data to
>>>> Pagecache than returned, although it called ::sync_file_range, but this
>>>> syscall doesn’t actually sync data to disk when it return,it’s an aync
>>>> call. So the situation is , the random write will be extremely fast
>>>> since it only write to journal and pagecache, but once syncing , it will
>>>> take very long time. The speed gap between journal and OSDs exist, the
>>>> amount of data that need to be sync keep increasing, and it will
>>>> certainly exceed 600s.
>>>>
>>>>
>>>>
>>>> For more information, I have tried to reproduce this by rados
>>>> bench,but failed.
>>>>
>>>>
>>>>
>>>> Could you please let me know if you need any more informations
>>>> & have some solutions? Thanks
>>>>
>>>>
>>>> Xiaoxi
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> [email protected]
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>
>>>
>>> --
>>> DI (FH) Wolfgang Hennerbichler
>>> Software Development
>>> Unit Advanced Computing Technologies
>>> RISC Software GmbH
>>> A company of the Johannes Kepler University Linz
>>>
>>> IT-Center
>>> Softwarepark 35
>>> 4232 Hagenberg
>>> Austria
>>>
>>> Phone: +43 7236 3343 245
>>> Fax: +43 7236 3343 250
>>> [email protected]
>>> http://www.risc-software.at
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> --
> DI (FH) Wolfgang Hennerbichler
> Software Development
> Unit Advanced Computing Technologies
> RISC Software GmbH
> A company of the Johannes Kepler University Linz
>
> IT-Center
> Softwarepark 35
> 4232 Hagenberg
> Austria
>
> Phone: +43 7236 3343 245
> Fax: +43 7236 3343 250
> [email protected]
> http://www.risc-software.at
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com