Need some help resolving the performance issues on the my ceph cluster. We are
running acute performance issues when we are using qemu-img convert. However
rbd import operation works perfectly alright. Please ignore image format for a
minute. I am trying to understand why rbd import performs well on the same
cluster where as qemu-img convert operation takes inordinate amount of time.
Here are the performance numbers:
1. qemu-img convert command for 465GB data took more than 48 hours to copy the
image to ceph.
[root@redhat-compute4 ~]# qemu-img convert -p -t none -O raw
/var/triliovault-mounts/MTAuMC4wLjc3Oi92YXIvbmZzX3NoYXJl/workload_326e8a43-a90a-4fe9-8aab-6d33bcdf5a05/snapshot_9f0cee13-8200-4562-82ec-1fb9f234bcd8/vm_id_05e9534e-5c84-4487-9613-1e0e227e4c1a/vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb/66582225-6539-4e5e-9b7a-59aa16739df1
rbd:vms/volume-5ad883a0cd65435bb6ffbfa1243bbdc6
(100.00/100%)
You have new mail in /var/spool/mail/root
[root@redhat-compute4 ~]#
2. Just copying the file to ceph took just 3 hours 18 mins (without qemu-img
convert).
[root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]# time
rbd import 66582225-6539-4e5e-9b7a-59aa16739df1 -p volumes
66582225-6539-4e5e-9b7a-59aa16739df1 --image-format 2
Importing image: 100% complete...done.
real 198m9.069s
user 5m32.724s
sys 18m32.213s
[root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
[root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]# rbd
info volumes/66582225-6539-4e5e-9b7a-59aa16739df1
rbd image '66582225-6539-4e5e-9b7a-59aa16739df1':
size 465 GB in 119081 objects
order 22 (4096 kB objects)
block_name_prefix: rbd_data.753102ae8944a
format: 2
features: layering
flags:
[root@redhat-compute4 vm_res_id_24291e4b-93d2-47ad-80a8-bf3c395319b9_vdb]#
I appreciate if any one can give me pointers on where to look for?
Best,
Murali Balcha
O 508.233.3912 | M 508.494.5007 | [email protected] | trilio.io
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com