Thanks for the suggestion, Simon.  Yes, we’ve looked at that, but we think that 
we’re going to potentially be in a situation where we’re using fairly big SSDs 
already.  For example, if we bought 30 6.4 TB SSDs rated at 1 DWPD and 
configured them as 6 4+1P RAID 5 LUNs, then we’d end up with a usable capacity 
of 6 * 4 * 6 = ~144 TB usable space in our “hot” pool.  That would satisfy our 
capacity needs and also not exceed the 1 DWPD rating of the drives.

BTW, we noticed with one particular vendor that their 3 DWPD drives were 
exactly 1/3rd the size of their 1 DWPD drives … which makes us wonder if that’s 
coincidence or not.  Anybody know for sure?

Thanks…

Kevin

> On Mar 18, 2019, at 4:13 PM, Simon Thompson <s.j.thomp...@bham.ac.uk> wrote:
> 
> Did you look at pricing larger SSDs than you need and only using partial 
> capacity to get more DWPD out of them?
> 
> I.e. 1TB drive 3dpwd = 3TBpd
> 2TB drive (using 1/2 capacity) = 6TBpd
> 
> Simon
> ________________________________________
> From: gpfsug-discuss-boun...@spectrumscale.org 
> [gpfsug-discuss-boun...@spectrumscale.org] on behalf of Buterbaugh, Kevin L 
> [kevin.buterba...@vanderbilt.edu]
> Sent: 18 March 2019 19:09
> To: gpfsug main discussion list
> Subject: Re: [gpfsug-discuss] SSDs for data - DWPD?
> 
> Hi All,
> 
> Just wanted to follow up with the results of my survey … I received a grand 
> total of two responses (Thanks Alex and John).  In their case, they’re using 
> SSDs with a 10 DWPD rating.
> 
> The motivation behind my asking this question was … money!  ;-).  Seriously, 
> 10 DWPD drives are still very expensive, while 3 DWPD drives are 
> significantly less expensive and 1 DWPD drives are even cheaper still.  While 
> we would NOT feel comfortable using anything less than 10 DWPD drives for 
> metadata, we’re wondering about using less expensive drives for data.
> 
> For example, let’s just say that you’re getting ready to set up a brand new 
> GPFS 5 formatted filesystem of 1-2 PB in size.  You’re considering having 3 
> pools:
> 
> 1) a metadata only system pool of 10 DWPD SSDs.  4K inodes, and a ton of 
> small files that’ll fit in the inode.
> 2) a data only “hot” pool (i.e. the default pool for writes) of SSDs.
> 3) a data only “capacity” pool of 12 TB spinning disks.
> 
> And let’s just say that you have looked back at the historical data you’ve 
> collected and you see that over the last 6 months or so you’ve been averaging 
> 10-12 TB of data being written into your existing filesystem per day.  You 
> want to do migrations between pools only on the weekends if at all possible.
> 
> 12 * 7 = 84 TB.  So if you had somewhere between 125 - 150 TB of SSDs ... 1 
> DWPD SSDs … then in theory you should easily be able to handle your 
> anticipated workload without coming close to exceeding the 1 DWPD rating of 
> the SSDs.
> 
> However, as the saying goes, while in theory there’s no difference between 
> theory and practice, in practice there is ... so am I overlooking anything 
> here from a GPFS perspective???
> 
> If anybody still wants to respond on the DWPD rating of the SSDs they use for 
> data, I’m still listening.
> 
> Thanks…
> 
> Kevin
> 
> P.S.  I still have a couple of “outstanding issues” to respond to that I’ve 
> posted to the list about previously:
> 
> 1) the long I/O’s we see occasionally in the output of “mmdiag —iohist” on 
> our NSD servers.  We’re still trying to track that down … it seems to happen 
> only with a subset of our hardware - most of the time at least - but we’re 
> still working to track down what triggers it … i.e. at this point I can’t say 
> whether it’s really the hardware or a user abusing the hardware.
> 
> 2) I promised to post benchmark results of 3 different metadata configs:  a) 
> RAID 1 mirrors, b) a RAID 5 stripe, c) no RAID, but GPFS metadata replication 
> of 3.  That benchmarking has been put on hold for reasons I can’t really 
> discuss on this mailing list at this time … but hopefully soon.
> 
> I haven’t forgotten the above and will respond back on the list when it’s 
> appropriate.  Thanks...
> 
> On Mar 8, 2019, at 10:24 AM, Buterbaugh, Kevin L 
> <kevin.buterba...@vanderbilt.edu<mailto:kevin.buterba...@vanderbilt.edu>> 
> wrote:
> 
> Hi All,
> 
> This is kind of a survey if you will, so for this one it might be best if you 
> responded directly to me and I’ll summarize the results next week.
> 
> Question 1 - do you use SSDs for data?  If not - i.e. if you only use SSDs 
> for metadata (as we currently do) - thanks, that’s all!  If, however, you do 
> use SSDs for data, please see Question 2.
> 
> Question 2 - what is the DWPD (daily writes per day) of the SSDs that you use 
> for data?
> 
> Question 3 - is that different than the DWPD of the SSDs for metadata?
> 
> Question 4 - any pertinent information in regards to your answers above (i.e. 
> if you’ve got a filesystem that data is uploaded to only once and never 
> modified after that then that’s useful to know!)?
> 
> Thanks…
> 
> Kevin
> 
> —
> Kevin Buterbaugh - Senior System Administrator
> Vanderbilt University - Advanced Computing Center for Research and Education
> kevin.buterba...@vanderbilt.edu<mailto:kevin.buterba...@vanderbilt.edu> - 
> (615)875-9633
> 
> 
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
> https://nam04.safelinks.protection.outlook.com/?url=http%3A%2F%2Fgpfsug.org%2Fmailman%2Flistinfo%2Fgpfsug-discuss&amp;data=02%7C01%7CKevin.Buterbaugh%40vanderbilt.edu%7C274d56e2906e4df3340a08d6abe6a61e%7Cba5a7f39e3be4ab3b45067fa80faecad%7C0%7C0%7C636885404456477052&amp;sdata=eJ6XKuMQ3H4y8V1kyTd8%2ByGJX0rhlTqfcl0fce14pYA%3D&amp;reserved=0

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to