Thanks for the work, Adam!

On 02/06/2013 03:18 PM, Michel, Adam K - (amichel) wrote:
I did some testing against an RBD device in my local environment to suss out 
differences between a few filesystem options. Folks in #ceph thought this data 
might be interesting to some on the list, so I'm posting a link to my 
spreadsheet here.

http://mirrors.arizona.edu/ceph-devel/oxide-rbd-benches.xlsx

My ceph cluster layout is pretty simple. I have one server node with all the 
OSDs. It's a backblaze-style pod with 40 drives, 1 OSD per drive, dual-core i3 
at 3GHz, 16GB of RAM and small SSD for system disk. It's running Ubuntu 12.04 
and attached with 10GbE to a switch in my lab. I have three mons on older Dell 
hardware, 2950s, dual quad-core E5345 Xeon, 32GB of RAM. They're running 
CentOS6 (needed support for the QL CNAs I had available) attached with 10GbE 
internal to the lab as well. I can provide ceph configuration data if anyone 
would like, but I followed the documentation pretty religiously so I expect it 
is quite vanilla. The client I used for testing is a VMware VM running Ubuntu 
12.04 in a different environment (ran out of hardware in the lab) which has a 
1GbE bottleneck in its path to the lab.

I created a 5TB rbd (it is theoretically going to end up in use in a pilot for 
a dropbox-style service for our campus, thus the rather strangely large size 
for testing) and mapped it to the client, created a GPT partition table with a 
single primary partition starting at sector 8192 (on guidance from someone in 
#ceph) and then formatted it with filesystem defaults for each of ext4, xfs and 
btrfs. I mounted with no special options in all cases.

I ran bonnie++ v1.97 with the option to skip char tests. For iozone I tested 
record sizes up to 4096 on default max file size of 512M. I've generated all 
the standard 3d charts for the iozone results in their respective sheets to the 
right of their matching data tables.

I make no particular claims to have done this as well as possible and would 
appreciate any feedback from the list on any kind of testing rubric that might 
generate data of more use to the community at large. This particular round of 
testing was mostly for my own edification.

Hopefully some of you find this useful!

Cheers,
Adam

--
Adam Michel
Systems Administrator
UITS, University of Arizona
[email protected]
5206262189

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to