With the low number of OSDs, you are probably satuarting the disks. Check
with `iostat -xd 2` and see what the utilization of your disks are. A lot
of SSDs don't perform well with Ceph's heavy sync writes and performance is
terrible.

If some of your drives are 100% while others are lower utilization, you can
possibly get more performance and greatly reduce the blocked I/O with the
WPQ scheduler. In the ceph.conf add this to the [osd] section and restart
the processes:

osd op queue = wpq
osd op queue cut off = high

This has helped our clusters with fairness between OSDs and making
backfills not so disruptive.
----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1


On Thu, Jun 6, 2019 at 1:43 AM BASSAGET Cédric <cedric.bassaget...@gmail.com>
wrote:

> Hello,
>
> I see messages related to REQUEST_SLOW a few times per day.
>
> here's my ceph -s  :
>
> root@ceph-pa2-1:/etc/ceph# ceph -s
>   cluster:
>     id:     72d94815-f057-4127-8914-448dfd25f5bc
>     health: HEALTH_OK
>
>   services:
>     mon: 3 daemons, quorum ceph-pa2-1,ceph-pa2-2,ceph-pa2-3
>     mgr: ceph-pa2-3(active), standbys: ceph-pa2-1, ceph-pa2-2
>     osd: 6 osds: 6 up, 6 in
>
>   data:
>     pools:   1 pools, 256 pgs
>     objects: 408.79k objects, 1.49TiB
>     usage:   4.44TiB used, 37.5TiB / 41.9TiB avail
>     pgs:     256 active+clean
>
>   io:
>     client:   8.00KiB/s rd, 17.2MiB/s wr, 1op/s rd, 546op/s wr
>
>
> Running ceph version 12.2.9 (9e300932ef8a8916fb3fda78c58691a6ab0f4217)
> luminous (stable)
>
> I've check :
> - all my network stack : OK ( 2*10G LAG )
> - memory usage : ok (256G on each host, about 2% used per osd)
> - cpu usage : OK (Intel(R) Xeon(R) CPU E5-2678 v3 @ 2.50GHz)
> - disk status : OK (SAMSUNG   AREA7680S5xnNTRI  3P04 => samsung DC series)
>
> I heard on IRC that it can be related to samsung PM / SM series.
>
> Do anybody here is facing the same problem ? What can I do to solve that ?
> Regards,
> Cédric
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to