Thank you. I was working in this direction. The situation is a lot better. But I think I can get still far better.
I could set the controller to writethrough, direct and no read ahead for the ssds. But I cannot disable the pdcache ☹ there is an option set in the controller "Block SSD Write Disk Cache Change = Yes" which does not permit to deactivate the ssd cache. I could not find any solution in google for this controller (LSI MegaRAID SAS 9271-8i) to change this setting. I don’t know how much performance gain it will be to deactivate the ssd cache. At least the micron 5200max has capacitor so I hope it is safe for data loss in case if power failure. I wrote a request to lsi / Broadcom if they know how I can change this setting. This is really annyoing. I will check the cpu power settings. I rode also somewhere it can improve iops a lot. (if its bad set) At the moment I get 600iops 4k random write 1 thread and 1 iodepth. I get 40K - 4k random iops for some instances with 32iodepth. Its not the world but a lot better then before. Read around 100k iops. For 16 ssd's and 2 x dual 10G nic. I was reading that good tunings and hardware config can get more then 2000 iops on single thread out of the ssds. I know thet ceph does not shine with single thread. But 600 iops is not very much... philipp -----Ursprüngliche Nachricht----- Von: Reed Dier <reed.d...@focusvq.com> Gesendet: Dienstag, 01. September 2020 22:37 An: VELARTIS Philipp Dürhammer <p.duerham...@velartis.at> Cc: ceph-users@ceph.io Betreff: Re: [ceph-users] Can 16 server grade ssd's be slower then 60 hdds? (no extra journals) If using storcli/perccli for manipulating the LSI controller, you can disable the on-disk write cache with: storcli /cx/vx set pdcache=off You can also ensure that you turn off write caching at the controller level with storcli /cx/vx set iopolicy=direct storcli /cx/vx set wrcache=wt You can also tweak the readahead value for the vd if you want, though with an ssd, I don't think it will be much of an issue. storcli /cx/vx set rdcache=nora I'm sure the megacli alternatives are available with some quick searches. May also want to check your c-states and p-states to make sure there isn't any aggressive power saving features getting in the way. Reed > On Aug 31, 2020, at 7:44 AM, VELARTIS Philipp Dürhammer > <p.duerham...@velartis.at> wrote: > > We have older LSi Raid controller with no HBA/JBOD option. So we expose the > single disks as raid0 devices. Ceph should not be aware of cache status? > But digging deeper in to it it seems that 1 out of 4 serves is performing a > lot better and has super low commit/applay rates while the other have a lot > mor (20+) on heavy writes. This just applys fore the ssd. For the hdds I cant > see a difference... > > -----Ursprüngliche Nachricht----- > Von: Frank Schilder <fr...@dtu.dk> > Gesendet: Montag, 31. August 2020 13:19 > An: VELARTIS Philipp Dürhammer <p.duerham...@velartis.at>; > 'ceph-users@ceph.io' <ceph-users@ceph.io> > Betreff: Re: Can 16 server grade ssd's be slower then 60 hdds? (no extra > journals) > > Yes, they can - if volatile write cache is not disabled. There are many > threads on this, also recent. Search for "disable write cache" and/or > "disable volatile write cache". > > You will also find different methods of doing this automatically. > > Best regards, > ================= > Frank Schilder > AIT Risø Campus > Bygning 109, rum S14 > > ________________________________________ > From: VELARTIS Philipp Dürhammer <p.duerham...@velartis.at> > Sent: 31 August 2020 13:02:45 > To: 'ceph-users@ceph.io' > Subject: [ceph-users] Can 16 server grade ssd's be slower then 60 hdds? (no > extra journals) > > I have a productive 60 osd's cluster. No extra Journals. Its performing okay. > Now I added an extra ssd Pool with 16 Micron 5100 MAX. And the performance is > little slower or equal to the 60 hdd pool. 4K random as also sequential > reads. All on dedicated 2 times 10G Network. HDDS are still on filestore. SSD > on bluestore. Ceph Luminous. > What should be possible 16 ssd's vs. 60 hhd's no extra journals? > > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to > ceph-users-le...@ceph.io > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io