On 11/07/2012 10:35 AM, Atchley, Scott wrote:
On Nov 7, 2012, at 11:20 AM, Mark Nelson <[email protected]> wrote:

  Right now I'm doing 3 journals per SSD, but topping out at about
1.2-1.4GB/s from the client perspective for the node with 15+ drives and
5 SSDs.  It's possible newer versions of the code and tuning may
increase that.

What interconnect is this? 10G Ethernet is 1.25 GB/s line rate and I would 
expect your Sockets and Ceph overhead to eat into that. Or is it dual 10G 
Ethernet?

This is 8 concurrent instances of rados bench running on localhost.
Ceph is configured with 1x replication.  1.2-1.4GB/s is the aggregate
throughput of all of the rados bench instances.

Ok, all local with no communication. Given this level of local performance, 
what does that translate into when talking over the network?

Scott


Well, local, but still over tcp. Right now I'm focusing on pushing the osds/filestores as far as I can, and after that I'm going to setup a bonded 10GbE network to see what kind of messenger bottlenecks I run into. Sadly the testing is going slower than I would like.

Mark
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to