Agreed.  I think oh …. maybe 15-20 years ago there was often a wider difference 
between SAS and SATA drives, but with modern queuing etc. my sense is that 
there is less of an advantage.  Seek and rotational latency I suspect dwarf 
interface differences wrt performance.  The HBA may be a bigger bottleneck (and 
way more trouble).

500 GB NVMe seems like a lot per HDD, are you using that as WAL+DB with RGW, or 
as dmcache or something?

Depending on your constraints, QLC flash might be more competitive than you 
think ;)

— aad


> I suspect the behavior of the controller and the behavior of the drive 
> firmware will end up mattering more than SAS vs SATA.  As always it's best if 
> you can test it first before committing to buying a pile of them.  
> Historically I have seen SATA drives that have performed well as far as HDDs 
> go though.
> 
> 
> Mark
> 
> On 6/3/21 4:25 PM, Dave Hall wrote:
>> Hello,
>> 
>> We're planning another batch of OSD nodes for our cluster.  Our prior nodes
>> have been 8 x 12TB SAS drives plus 500GB NVMe per HDD.  Due to market
>> circumstances and the shortage of drives those 12TB SAS drives are in short
>> supply.
>> 
>> Our integrator has offered an option of 8 x 14TB SATA drives (still
>> Enterprise).  For Ceph, will the switch to SATA carry a performance
>> difference that I should be concerned about?
>> 
>> Thanks.
>> 
>> -Dave
>> 
>> --
>> Dave Hall
>> Binghamton University
>> kdh...@binghamton.edu
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to