> So my question is what is the use case. S3 is not a use case it’s a protocol 
> to access the storage.

Indeed.  Factors include:
* read vs writes, many RGW deployments are very read-heavy
* distribution of object sizes.  Smaller objects (per-second checkpoints) are 
much more metadata-intensive than larger objects (pirated TV shows dubbed into 
Portuguese).

> How many nodes are you looking to deploy?

I suggest that a chassis not hold more than 10% of the cluster capacity with 
slow media, for blast radius.

> What model of Kioxia drive are you looking to use as a brief look said all 
> the 30TB drives where read intensive.

Be very sure that those are TLC, 4KB IU drives, as I believe they have made 
30TB QLC drives as well, which most likely would not be ideal for this purpose. 
 But would be for large-object bucket data pools.

> If you go with 90 drive chassis then you are looking at a ratio of 1:22 for 
> NVME to HDD

Indeed, conventional wisdom has been at most a 10:1 ratio, though with modern 
PCIe 4+ servers and drives, a somewhat higher ratio is probably tolerable.  But 
those would also want to have OSDs for the meta, log, etc pools.

> which could be a big issue made worse by the use of read intensive drives 
> that may only do 80K write IOPS. Back to the first question.

Remember that mixed-use drives are usually the same hardware with more 
overprovisioning (and profit margin).

> Last one is where are you going to put the S3 bucket metadata?

The index pool is mainly omaps (so far), which can live on hybrid OSDs with SSD 
offload.

> Comes back too the first question of what is the use case ?
> 
> Like Mark one of the configurations I use a lot is the 2U24. Whilst the 90 
> drive chassis looks good on paper it typically needs a deeper rack is very 
> heavy and also consumes a lot of power. You are now at close to 2PB per node 
> which is a massive amount of data to rebuild should a node be lost. Unless 
> you are wanting just a very deep archive and a cluster of 30+PB then IMHO 
> they don’t make sense.

Agreed.

> 
> 
>> On 7 Aug 2025, at 17:54, Mark Nelson <mark.nel...@clyso.com> wrote:
>> 
>> Personally I prefer the 24 HDD + 4NVMe 2U chassis unless extreme density and 
>> absolute lowest cost are the highest priorities.  The denser options can 
>> work, but definitely have trade-offs and are only really appropriate if you 
>> are deploying a very large cluster.
>> 
>> 
>> Mark
>> 
>> 
>> On 8/7/25 09:48, Manuel Rios - EDH wrote:
>>> 
>>> Hi Ceph ,
>>> 
>>> Does anyone deployed CEPH in supermicro super storage nodes of 60/90 HDD + 
>>> 4 NVME for WALL?
>>> 
>>> Newest models support 144cores in single socket and several TB of ram 
>>> without issues.
>>> 
>>> But as far as we understood in the technical notes it use SAS Expanders for 
>>> connect all disk, from 2 to 4 SAS Expanders to connect the whole chasis.
>>> 
>>> We’re looking for next configuration :
>>> 
>>> Intel Xeon 96cores @ 2.4
>>> 
>>> 1TB RAM
>>> 
>>> 2 SSD for OS.
>>> 
>>> 4 NVME x 30TB NVME KIOXIA
>>> 
>>> 90 HDD * 22 TB HGST
>>> 
>>> 4x25 Gbps or 2x100 Gbps
>>> 
>>> Main use RGW.
>>> 
>>> Regards
>>> 
>>>     
>>> 
>>> *MANUEL RIOS FERNANDEZ*
>>> 
>>> CEO –  EasyDataHost
>>> 
>>> *Phone: * 677677179
>>> 
>>> *Web: *_www.easydatahost.com <http://www.easydatahost.com/>_
>>> 
>>> *Email: *_mrios...@easydatahost.com <mailto:mrios...@easydatahost.com>_
>>> 
>>> Título: LinkedIn - Descripción: image of LinkedIn icon 
>>> <https://es.linkedin.com/in/manuel-rios-fernandez-14880949?original_referer=https%3A%2F%2Fwww.google.com%2F>
>>> 
>>> *_ADVERTENCIA LEGAL:_*
>>> 
>>> Este mensaje y, en su caso, los ficheros anexos son confidenciales, 
>>> especialmente en lo que respecta a los datos personales, y se dirigen 
>>> exclusivamente al destinatario referenciado.
>>> 
>>> Si usted no lo es y lo ha recibido por error o tiene conocimiento del mismo 
>>> por cualquier motivo, le rogamos que nos lo comunique por este medio y 
>>> proceda a destruirlo o borrarlo, y que en todo caso se abstenga de 
>>> utilizar, reproducir, alterar, archivar o comunicar a terceros el presente 
>>> mensaje y ficheros anexos, todo ello bajo pena de incurrir en 
>>> responsabilidades legales. El emisor no garantiza la integridad, rapidez o 
>>> seguridad del presente correo, ni se responsabiliza de posibles perjuicios 
>>> derivados de la captura, incorporaciones de virus o cualesquiera otras 
>>> manipulaciones efectuadas por terceros.
>>> 
>>> *_CONFIDENTIALITY NOTICE:_*
>>> 
>>> This e-mail message and all attachments transmitted with it may contain 
>>> legally privileged, proprietary and/or confidential information intended 
>>> solely for the use of the addressee. If you are not the intended recipient, 
>>> you are hereby notified that any review, dissemination, distribution, 
>>> duplication or other use of this message and/or its attachments is strictly 
>>> prohibited. If you are not the intended recipient, please contact the 
>>> sender by reply e-mail and destroy all copies of the original message and 
>>> its attachments. Thank you.
>>> 
>>> *Descripción: Descripción: Descripción: Descripción: Descripción: Flor 
>>> ecoTECHNo imprimas si no es necesario. Protejamos el Medio Ambiente.*
>>> 
>>> 
>>> _______________________________________________
>>> ceph-users mailing list --ceph-users@ceph.io
>>> To unsubscribe send an email toceph-users-le...@ceph.io
>> 
>> -- 
>> Best Regards,
>> Mark Nelson
>> Head of Research and Development
>> 
>> Clyso GmbH
>> p: +49 89 21552391 12 | a: Minnesota, USA
>> w:https://clyso.com  | e:mark.nel...@clyso.com
>> 
>> We are hiring:https://www.clyso.com/jobs/
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to