> On Sep 5, 2018, at 11:34 AM, Buterbaugh, Kevin L 
> <[email protected] <mailto:[email protected]>> 
> wrote:
> 
> 

[…]

> Of course, if we increase the size of the inodes by a factor of 8 then we 
> also need 8 times as much space to store those inodes.  Given that Enterprise 
> class SSDs are still very expensive and our budget is not unlimited, we’re 
> trying to get the best bang for the buck.
> 

Nobody has gone in this direction yet, so I’ll play devil’s advocate:

Are you sure you need enterprise class SSDs? The only practical difference 
between enterprise class SSDs and "read intensive" SSDs is the "endurance" in 
DWPD[1]. Read-intensive SSDs usually have a DWPD of 1-ish. Enterprise SSDs can 
have a DWPD as high as 30.

So, how many times do you think you’ll actually write all of the data on the 
SSDs per day?

I don’t know how much (meta)data you’ve got, but maybe consider buying the 
"cheap" SSDs (which will be *much* larger for your dollar) and just use 
fractions of them with GPFS replication[2] or maybe some vendor’s {distributed, 
de-clustererd} RAID. Keep some spares.

This is probably bad advice, but the thought exercise will let you find the 
edges of what you meant. :)

[1] DWPD = Drive Writes Per Day — write all of the cells on the entire storage 
device every 24 hours.
[2] Okay, somebody already said to use GPFS replication. ;)

-- 
Stephen



> We have always - even back in the day when our metadata was on spinning disk 
> and not SSD - used RAID 1 mirrors and metadata replication of two.  However, 
> we are wondering if it might be possible to switch to RAID 5?  Specifically, 
> what we are considering doing is buying 8 new SSDs and creating two 3+1P RAID 
> 5 LUNs (metadata replication would stay at two).  That would give us 50% more 
> usable space than if we configured those same 8 drives as four RAID 1 mirrors.
> 
> Unfortunately, unless I’m misunderstanding something, mean that the RAID 
> stripe size and the GPFS block size could not match.  Therefore, even though 
> we don’t need the space, would we be much better off to buy 10 SSDs and 
> create two 4+1P RAID 5 LUNs?
> 
> I’ve searched the mailing list archives and scanned the DeveloperWorks wiki 
> and even glanced at the GPFS documentation and haven’t found anything that 
> says “bad idea, Kevin”… ;-)
> 
> Expanding on this further … if we just present those two RAID 5 LUNs to GPFS 
> as NSDs then we can only have two NSD servers as primary for them.  So 
> another thing we’re considering is to take those RAID 5 LUNs and further 
> sub-divide them into a total of 8 logical volumes, each of which could be a 
> GPFS NSD and therefore would allow us to have each of our 8 NSD servers be 
> primary for one of them.  Even worse idea?!?  Good idea?
> 
> Anybody have any better ideas???  ;-)
> 
> Oh, and currently we’re on GPFS 4.2.3-10, but are also planning on moving to 
> GPFS 5.0.1-x before creating the new filesystem.
> 
> Thanks much…
> 
> —
> Kevin Buterbaugh - Senior System Administrator
> Vanderbilt University - Advanced Computing Center for Research and Education
> [email protected] <mailto:[email protected]> - 
> (615)875-9633
> 
> 
> 
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org <http://spectrumscale.org/>
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss 
> <http://gpfsug.org/mailman/listinfo/gpfsug-discuss>

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to