hope this helpful. http://software.intel.com/en-us/blogs/2013/10/25/measure-ceph-rbd-performance-in-a-quantitative-way-part-i http://software.intel.com/en-us/blogs/2013/11/20/measure-ceph-rbd-performance-in-a-quantitative-way-part-ii
On Thu, Dec 12, 2013 at 4:22 PM, Wido den Hollander <[email protected]> wrote: > On 12/11/2013 09:13 PM, German Anders wrote: > >> Hi to all, >> >> I'm new to Ceph and i want to create a Cluster for Production >> with HP ProLiant DL380p Gen8 servers, the idea is to use 4 of this >> servers to use as OSD's, and then 3 x HP ProLiant DL320e Gen8 servers >> for MON. The Data network would be on 10GbE switches and 1Gb the >> management. Below is the description of each of the servers: >> >> > In this case I would run the monitors on the same machines, since a DL320 > is very overpowered for a monitor in this cluster set up. > > *HP ProLiant DL380p Gen8*: >> 2 x Intel Xeon E5-2630v2 @2.6Ghz (6-cores) >> 2 x 64GB RAM >> 2 x 450GB SAS 15K in RAID-1 configuration for the OS >> 2 x 100GB SSD in RAID-1 configuration for the Journals >> > > I wouldn't use RAID-1. Short version: SSDs rarely fail and if they fail it > is due to wearing out. If you use RAID-1 they'll fail at the same moment. > > You better use one SSD per 4 OSDs, gives you better performance and > reliability. > > 8 x 4TB SATA 7.2K to use as 8 x OSD's (32TB raw) >> 1 x HP Ethernet 10GbE 2-port 530SFP+ Adapter >> 1 x HP Ethernet 1Gb 2-port 332T Adapter >> >> *HP ProLiant DL320e Gen8*: >> 1 x Intel Xeon E3-1240v2 @3.4Ghz (4-cores) >> 1 x 32GB RAM >> > > Way to much memory for a monitor. 4GB ~ 8GB is more then enough. > > 2 x 450GB SAS 15K in RAID-1 configuration for the OS >> 2 x 1.2TB SAS 10K for Logs >> 1 x HP Ethernet 10GbE 2-port 530SFP+ Adapter >> > > 10Gbit isn't required, but that's up to you. > > 1 x HP Ethernet 1Gb 2-port 332T Adapter >> >> >> I want to know if someone more or less had a similar configuration and >> to know what are the performance numbers (some benchmarks) for reads and >> writes, maybe also some iozone or bonnie++ outputs, with several >> processes (1..10), and different block sizes. >> Also if anybody had some recommendations or tips regarding the >> configuration for performance. The filesystem to be used is XFS. >> >> > I assume you are going for 3x replication, so with writes you'll have > about 1/3 of the I/O performance of all the disks. > > A 7200RPM disk is capable of about 100 IOps, so that's the figure you > calculate with. > > Ceph performance is very complex, so one bonnie++ or iozone benchmark > won't reflect the performance of a other Ceph setup. > > Wido > > I really appreciated the help. >> >> Thanks in advance, >> >> Best regards, >> >> *German Anders* >> >> >> >> >> >> >> >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > -- > Wido den Hollander > 42on B.V. > > Phone: +31 (0)20 700 9902 > Skype: contact42on > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
