Hi German,

 

Are the power feeds completely separate (ie 4 feeds in total), or just each 
rack has both feeds? If it’s the latter I don’t see any benefit from including 
this into the crushmap and would just create a “rack” bucket. Also assuming 
your servers have dual PSU’s, this also changes the power failure scenarios 
quite a bit as well.

 

In regards to the pools, unless you know your workload will easily fit into a 
cache pool with room to spare, I would suggest not going down that route 
currently. Performance in many cases can actually end up being worse if you end 
up doing a lot of promotions.

 

*However* I’ve been doing a bit of testing with the current master and there 
are a lot of improvements around cache tiering that are starting to have a 
massive improvement on performance. If you can get by with just the SAS disks 
for now and make a more informed decision about the cache tiering when 
Infernalis is released then that might be your best bet.

 

Otherwise you might just be best using them as a basic SSD only Pool.

 

Nick

 

From: ceph-users [mailto:[email protected]] On Behalf Of German 
Anders
Sent: 04 September 2015 16:30
To: ceph-users <[email protected]>
Subject: [ceph-users] Best layout for SSD & SAS OSDs

 

Hi cephers,

   I've the following scheme:

7x OSD servers with:

    4x 800GB SSD Intel DC S3510 (OSD-SSD)

    3x 120GB SSD Intel DC S3500 (Journals)

    5x 3TB SAS disks (OSD-SAS)

The OSD servers are located on two separate Racks with two power circuits each.

   I would like to know what is the best way to implement this.. use the 4x 
800GB SSD like a SSD-pool, or used them us a Cache pool? or any other 
suggestion? Also any advice for the crush design?

Thanks in advance,    




German




_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to