Re: [ceph-users] Tuning Nautilus for flash only

2019-11-28 Thread Paul Emmerich
Can confirm that disabling power saving helps. I've also seen latency
improvements with sysctl -w net.ipv4.tcp_low_latency=1

Another thing that sometimes helps is disabling the write cache of
your SSDs (hdparm -W 0), depends on the disk model, though.

Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

On Thu, Nov 28, 2019 at 7:59 PM David Majchrzak, ODERLAND Webbhotell
AB  wrote:
>
> Paul,
>
> Absolutely, I said I was looking at those settings and most didn't make
> any sense to me in a production environment (we've been running ceph
> since Dumpling).
>
> However we only have 1 cluster on Bluestore and I wanted to get some
> opinions if anything other than the defaults in ceph.conf or sysctl or
> things like Wido suggested with c-states would make any differences.
> (Thank you Wido!)
>
> Yes, running benchmarks is great, and we're already doing that
> ourselves.
>
> Cheers and have a nice evening!
>
> --
> David Majchrzak
>
>
> On tor, 2019-11-28 at 17:46 +0100, Paul Emmerich wrote:
> > Please don't run this config in production.
> > Disabling checksumming is a bad idea, disabling authentication is
> > also
> > pretty bad.
> >
> > There are also a few options in there that no longer exist (osd op
> > threads) or are no longer relevant (max open files), in general, you
> > should not blindly copy config files you find on the Internet. Only
> > set an option to its non-default value after carefully checking what
> > it does and whether it applies to your use case.
> >
> > Also, run benchmarks yourself. Use benchmarks that are relevant to
> > your use case.
> >
> > Paul
> >
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Tuning Nautilus for flash only

2019-11-28 Thread David Majchrzak, ODERLAND Webbhotell AB
Paul,

Absolutely, I said I was looking at those settings and most didn't make
any sense to me in a production environment (we've been running ceph
since Dumpling).

However we only have 1 cluster on Bluestore and I wanted to get some
opinions if anything other than the defaults in ceph.conf or sysctl or
things like Wido suggested with c-states would make any differences.
(Thank you Wido!)

Yes, running benchmarks is great, and we're already doing that
ourselves.

Cheers and have a nice evening!

-- 
David Majchrzak


On tor, 2019-11-28 at 17:46 +0100, Paul Emmerich wrote:
> Please don't run this config in production.
> Disabling checksumming is a bad idea, disabling authentication is
> also
> pretty bad.
> 
> There are also a few options in there that no longer exist (osd op
> threads) or are no longer relevant (max open files), in general, you
> should not blindly copy config files you find on the Internet. Only
> set an option to its non-default value after carefully checking what
> it does and whether it applies to your use case.
> 
> Also, run benchmarks yourself. Use benchmarks that are relevant to
> your use case.
> 
> Paul
> 

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Tuning Nautilus for flash only

2019-11-28 Thread Paul Emmerich
Please don't run this config in production.
Disabling checksumming is a bad idea, disabling authentication is also
pretty bad.

There are also a few options in there that no longer exist (osd op
threads) or are no longer relevant (max open files), in general, you
should not blindly copy config files you find on the Internet. Only
set an option to its non-default value after carefully checking what
it does and whether it applies to your use case.

Also, run benchmarks yourself. Use benchmarks that are relevant to
your use case.

Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

On Thu, Nov 28, 2019 at 1:17 PM Wido den Hollander  wrote:
>
>
>
> On 11/28/19 12:56 PM, David Majchrzak, ODERLAND Webbhotell AB wrote:
> > Hi!
> >
> > We've deployed a new flash only ceph cluster running Nautilus and I'm
> > currently looking at any tunables we should set to get the most out of
> > our NVMe SSDs.
> >
> > I've been looking a bit at the options from the blog post here:
> >
> > https://ceph.io/community/bluestore-default-vs-tuned-performance-comparison/
> >
> > with the conf here:
> > https://gist.github.com/likid0/1b52631ff5d0d649a22a3f30106ccea7
> >
> > However some of them, like checksumming, is for testing speed only but
> > not really applicable in a real life scenario with critical data.
> >
> > Should we stick with defaults or is there anything that could help?
> >
> > We have 256GB of RAM on each OSD host, 8 OSD hosts with 10 SSDs on
> > each. 2 osd daemons on each SSD. Raise ssd bluestore cache to 8GB?
> >
> > Workload is about 50/50 r/w ops running qemu VMs through librbd. So
> > mixed block size.
>
> Pin the C-State of your CPUs to 1 and disable powersaving. That can
> reduce the latency vastly.
>
> Testing with rados bench -t 1 -b 4096 -o 4096 you should be able to get
> to a 0.8ms write latency with 3x replication.
>
> >
> > 3 replicas.
> >
> > Appreciate any advice!
> >
> > Kind Regards,
> >
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Tuning Nautilus for flash only

2019-11-28 Thread Wido den Hollander



On 11/28/19 12:56 PM, David Majchrzak, ODERLAND Webbhotell AB wrote:
> Hi!
> 
> We've deployed a new flash only ceph cluster running Nautilus and I'm
> currently looking at any tunables we should set to get the most out of
> our NVMe SSDs.
> 
> I've been looking a bit at the options from the blog post here:
> 
> https://ceph.io/community/bluestore-default-vs-tuned-performance-comparison/
> 
> with the conf here:
> https://gist.github.com/likid0/1b52631ff5d0d649a22a3f30106ccea7
> 
> However some of them, like checksumming, is for testing speed only but
> not really applicable in a real life scenario with critical data.
> 
> Should we stick with defaults or is there anything that could help?
> 
> We have 256GB of RAM on each OSD host, 8 OSD hosts with 10 SSDs on
> each. 2 osd daemons on each SSD. Raise ssd bluestore cache to 8GB?
> 
> Workload is about 50/50 r/w ops running qemu VMs through librbd. So
> mixed block size.

Pin the C-State of your CPUs to 1 and disable powersaving. That can
reduce the latency vastly.

Testing with rados bench -t 1 -b 4096 -o 4096 you should be able to get
to a 0.8ms write latency with 3x replication.

> 
> 3 replicas.
> 
> Appreciate any advice!
> 
> Kind Regards,
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Tuning Nautilus for flash only

2019-11-28 Thread David Majchrzak, ODERLAND Webbhotell AB
Hi!

We've deployed a new flash only ceph cluster running Nautilus and I'm
currently looking at any tunables we should set to get the most out of
our NVMe SSDs.

I've been looking a bit at the options from the blog post here:

https://ceph.io/community/bluestore-default-vs-tuned-performance-comparison/

with the conf here:
https://gist.github.com/likid0/1b52631ff5d0d649a22a3f30106ccea7

However some of them, like checksumming, is for testing speed only but
not really applicable in a real life scenario with critical data.

Should we stick with defaults or is there anything that could help?

We have 256GB of RAM on each OSD host, 8 OSD hosts with 10 SSDs on
each. 2 osd daemons on each SSD. Raise ssd bluestore cache to 8GB?

Workload is about 50/50 r/w ops running qemu VMs through librbd. So
mixed block size.

3 replicas.

Appreciate any advice!

Kind Regards,
-- 
David Majchrzak


___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com