Perhaps just one cluster has low latency and the other has excessively
high latency? You can use "rbd bench-write" to verify.

On Wed, Jun 28, 2017 at 8:04 PM, Murali Balcha <[email protected]> wrote:
> We will give it a try. I have another cluster of similar configuration and 
> the converts are working fine. We have not changed  any queue depth setting 
> on that setup either. If it turns out to be queue depth how can we set queue 
> setting  for qemu-img convert operation?
>
> Thank you.
>
> Sent from my iPhone
>
>> On Jun 28, 2017, at 7:56 PM, Jason Dillaman <[email protected]> wrote:
>>
>> Given that your time difference is roughly 10x, best guess is that
>> qemu-img is sending the IO operations synchronously (queue depth = 1),
>> whereas, by default, "rbd import" will send up to 10 write requests in
>> parallel to the backing OSDs. Such an assumption assumes that you have
>> really high latency. You can re-run "rbd import" with
>> "--rbd-concurrent-management-ops=1" to change your queue depth to 1
>> and see if it's similar to qemu-img runtime.
>>
>>> On Wed, Jun 28, 2017 at 5:46 PM, Murali Balcha <[email protected]> 
>>> wrote:
>>> Need some help resolving the performance issues on the my ceph cluster. We
>>> are running acute performance issues when we are using qemu-img convert.
>>> However rbd import operation works perfectly alright. Please ignore image
>>> format for a minute. I am trying to understand why rbd import performs well
>>> on the same cluster where as qemu-img convert operation takes inordinate
>>> amount of time. Here are the performance numbers:
>>>
>>> 1. qemu-img convert command for 465GB data took more than 48 hours to copy
>>> the image to ceph.
>>>
>>> [root@redhat-compute4 ~]# qemu-img convert -p -t none -O raw
>>> /var/triliovault-mounts/MTAuMC4wLjc3Oi92YXIvbmZzX3NoYXJl/workload_326e8a43-a90a-4fe9-8aab-6d33bcdf5a05/snapshot_9f0cee13-8200-4562-82ec-1fb9f234bcd8/vm_id_05e9534e-5c84-4487-9613-1e0e227e4c1a/vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb/66582225-6539-4e5e-9b7a-59aa16739df1
>>> rbd:vms/volume-5ad883a0cd65435bb6ffbfa1243bbdc6
>>>
>>>    (100.00/100%)
>>>
>>> You have new mail in /var/spool/mail/root
>>>
>>> [root@redhat-compute4 ~]#
>>>
>>>
>>> 2. Just copying the file to ceph took just 3 hours 18 mins (without qemu-img
>>> convert).
>>>
>>> [root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
>>> time rbd import 66582225-6539-4e5e-9b7a-59aa16739df1 -p volumes
>>> 66582225-6539-4e5e-9b7a-59aa16739df1 --image-format 2
>>>
>>> Importing image: 100% complete...done.
>>>
>>>
>>> real    198m9.069s
>>>
>>> user    5m32.724s
>>>
>>> sys     18m32.213s
>>>
>>> [root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
>>>
>>> [root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
>>> rbd info volumes/66582225-6539-4e5e-9b7a-59aa16739df1
>>>
>>> rbd image '66582225-6539-4e5e-9b7a-59aa16739df1':
>>>
>>>        size 465 GB in 119081 objects
>>>
>>>        order 22 (4096 kB objects)
>>>
>>>        block_name_prefix: rbd_data.753102ae8944a
>>>
>>>        format: 2
>>>
>>>        features: layering
>>>
>>>        flags:
>>>
>>> [root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
>>>
>>>
>>> I appreciate if any one can give me pointers on where to look for?
>>>
>>> Best,
>>>
>>> Murali Balcha
>>> O 508.233.3912 | M 508.494.5007 | [email protected] | trilio.io
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> [email protected]
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>>
>>
>> --
>> Jason



-- 
Jason
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to