In that case you need fair IOPS and high throughput. Go with S3610 or the 
Samsungs (or something else that people there can recommend, but for the love 
of god don't save on drives :)). It's easier to stick to one type of drives and 
not complicate things.

I would also recommend you add one storage node and make the CRUSH with 
intermediate level for groups of 3 nodes. Makes maintenance much easier and the 
data more durable in case of failure. (Best to put the nodes in separate cages 
on different UPSes, then you can do stuff like disable barriers if you go with 
some cheaper drives that need it.) I'm not a CRUSH expert, there are more 
tricks to do before you set this up.

Jan

> On 27 Aug 2015, at 18:31, German Anders <[email protected]> wrote:
> 
> Hi Jan,
> 
>    Thanks for responding the email, regarding the cluster usage, we are going 
> to used it for non-relational databases, Cassandra, mongoDBs and other apps, 
> so we need that this cluster response well to intense io apps, it's going to 
> be connected to HP enclosures with IB FDR also, and mapped through Cinder to 
> mount it on VMs (KVM hypervisor), then on the vms we are going to run the 
> non-relational dbs.
> 
> Thanks in advance,
> 
> 
> 
> German
> 
> 2015-08-27 13:25 GMT-03:00 Jan Schermer <[email protected] 
> <mailto:[email protected]>>:
> Some comments inline.
> A lot of it depends on your workload, but I'd say you almost certainly need 
> higher-grade SSDs. You can save money on memory.
> 
> What will be the role of this cluster? VM disks? Object storage? Streaming?...
> 
> Jan
> 
>> On 27 Aug 2015, at 17:56, German Anders <[email protected] 
>> <mailto:[email protected]>> wrote:
>> 
>> Hi all,
>> 
>>    I'm planning to deploy a new Ceph cluster with IB FDR 56Gb/s and I've the 
>> following HW:
>> 
>> 3x MON Servers:
>>    2x Intel Xeon E5-2600@v3 8C
>>    256GB RAM
> 
> I don't think you need that much memory, 64GB should be plenty (if that's the 
> only role for the servers).
> 
>>    1xIB FRD ADPT-DP (two ports for PUB network)
>>    1xGB ADPT-DP
>>    
>>    Disk Layout:
>>    
>>    SOFT-RAID:
>>    SCSI1 (0,0,0) (sda) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
>>    SCSI2 (0,0,0) (sdb) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
> 
> I 100% recommend going with SSDs for the /var/lib/ceph/mon storage, fast ones 
> (but they can be fairly small). Should be the same grade as journal drives 
> IMO.
> NOT S3500!
> I can recommend S3610 (just got some :)), Samsung 845 DC PRO. At least 1 DWPD 
> rating, better go with 3 DWPD.
> 
>> 
>> 8x OSD Servers:
>>    2x Intel Xeon E5-2600@v3 10C
> 
> Go for the fastest you can afford if you need the latency - even at the 
> expense of cores.
> Go for cores if you want bigger throughput.
> 
>>    256GB RAM
> 
> Again - I think too much if that's the only role for those nodes, 64GB should 
> be plenty.
> 
> 
>>    1xIB FRD ADPT-DP (one port for PUB and one for CLUS network)
>>    1xGB ADPT-DP
>> 
>>    Disk Layout:
>> 
>>    SOFT-RAID:
>>    SCSI1 (0,0,0) (sda) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
>>    SCSI2 (0,0,0) (sdb) - 120.0 GB ATA INTEL SSDSC2BB12 (OS-RAID1)
>> 
>>    JBOD:
>>    SCSI9 (0,0,0) (sdd) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)
>>    SCSI9 (0,1,0) (sde) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)
>>    SCSI9 (0,2,0) (sdf) - 120.0 GB ATA INTEL SC3500 SSDSC2BB12 (Journal)
> 
> No no no. Those SSDs will die a horrible death, too little endurance.
> Better go with 2x 3700 in RAID1 and partition them for journals. Or just 
> don't use journaling drives and buy better SSDs for storage.
> 
>>    SCSI9 (0,3,0) (sdg) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD) 
>>    SCSI9 (0,4,0) (sdh) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
>>    SCSI9 (0,5,0) (sdi) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
>>    SCSI9 (0,6,0) (sdj) - 800.2 GB ATA INTEL SC3510 SSDSC2BB80 (Pool-SSD)
> 
> Too little endurance.
> 
>> 
>>    SCSI9 (0,7,0) (sdk) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA) 
>>    SCSI9 (0,8,0) (sdl) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
>>    SCSI9 (0,9,0) (sdm) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
>>    SCSI9 (0,10,0) (sdn) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
>>    SCSI9 (0,11,0) (sdo) - 3.0 TB SEAGATE ST3000NM0023 (Pool-SATA)
>> 
>> 
>> I would like to have an expert opinion on what would be the best 
>> deploy/config disk pools and crush map? any other advice?
>> 
>> Thanks in advance,
>> 
>> Best regards,
>> 
>> 
>> German
>> _______________________________________________
>> ceph-users mailing list
>> [email protected] <mailto:[email protected]>
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
> 
> 

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to