hi,
I have a few question about the rbd performance. I have a small ceph installation, three osd server one monitor server and one compute node which maps a rbd image to a block device, all server a connectet via a dedicated 1Gbs network.
Each osd is capable of doing around 90MB/s tested with osd bench.
But if I test the write speed of the rbd block device the performance ist quite poor.

I do the test with
dd if=/dev/zero of=/dev/rbd0 bs=1M count=10000 oflag=direct,
I get a throughput around 25MB/s.
I used wireshark to graph the network throughput, the image is
http://tuxadero.com/multistorage/ceph.jpg
as you can see the throughput is not smooth.

The graph for the test without the oflag=direct is
http://tuxadero.com/multistorage/ceph2.jpg
which is much better, but I the compute node uses around 4-5G of it's RAM as a writeback cache, which is not acceptable for my application.

For comparison the graph for a scp transfer.
http://tuxadero.com/multistorage/scp.jpg

I read in the ceph doku, that ever "package" has to be commited to the disk on the osd, before it is acknowledged to the client, could you please expalain what a package is? Probably not a TCP package.

And on the mailinglist was a discussion about a writeback window, to my understanding it say how many byte can be
unacknowledged in transit, is that right?

How could I activate it?

Thanks for your time.

Best Regards,
 martin
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to