We tested the m500 960GB for journaling and found at most it could journal
3 spinner OSDs. I'd strongly recommend you avoid the Crucial consumer
drives based on our testing/usage. We ended up journaling those to the
spinner itself and getting better performance. Also, I wouldn't trust their
power loss protection and would assume a host is dead if it ever powers
down unexpectedly with those as journal devices.

On Tue, Apr 28, 2015 at 5:34 AM, Dominik Hannen <[email protected]> wrote:

> Hi ceph-users,
>
> I am currently planning a cluster and would like some input specifically
> about the storage-nodes.
>
> The non-osd systems will be running on more powerful system.
>
> Interconnect as currently planned:
> 4 x 1Gbit LACP Bonds over a pair of MLAG-capable switches (planned: EX3300)
>
> So far I would go with Supermicros 5018A-MHN4 offering, rack-space is not
> really a concern, so only 4 OSDs per U is fine.
> (The cluster is planned to start with 8 osd-nodes.)
>
> osd-node:
> Avoton C2758 - 8 x 2.40GHz
> 16 GB RAM ECC
> 16 GB SSD - OS - SATA-DOM
> 250GB SSD - Journal (MX200 250GB with extreme over-provisioning, staggered
> deployment, monitored for TBW-Value)
> 4 x 3 TB OSD - Seagate Surveillance HDD (ST3000VX000) 7200rpm 24/7
> 4 x 1 Gbit
>
> per-osd breakdown:
> 3 TB HDD
> 2 x 2.40GHz (Avoton-Cores)
> 4 GB RAM
> 8 GB SSD-Journal (~125 MB/s r/w)
> 1 Gbit
>
> The main question is, will the Avoton CPU suffice? (I recon the common
> 1GHz/OSD suggestion are in regards to much more powerful CPUs.)
>
> Are there any cost-effective suggestions to improve this configuration?
>
> Will erasure coding be a feasible possibility?
>
> Does it hurt to run OSD-nodes CPU-capped, if you have enough of them?
>
> ___
> Dominik Hannen
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



-- 
David Burley
NOC Manager, Sr. Systems Programmer/Analyst
Slashdot Media

e: [email protected]
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to