Hi,
I have all the normal options set in ceph.conf (priority and class for disk
threads) however scrubs look like they are running as
the standard BE/4 class in iotop. Running 10.2.3.
Eg
PG Dump (Shows that OSD 1 will be scrubbing)
pg_stat objects mip degr misp unf bytes log disklog state
state_stamp v reported up
up_primary acting acting_primary last_scrub scrub_stamp
last_deep_scrub deep_scrub_stamp
1.8fc 3314 0 0 0 0 13899923456 3043 3043
active+clean+scrubbing+deep 2016-11-08
07:51:25.708169 1468442'173537 1468443:139543 [15,32,1] 15
[15,32,1] 15 1435301'169725 2016-11-06
21:46:26.537632 1338333'158008 2016-10-30 23:59:50.794774
sudo iotop -o --batch --iter 1 | grep ceph
3958453 be/4 ceph 92.17 M/s 0.00 B/s 0.00 % 66.59 % ceph-osd -f
--cluster ceph --id 1 --setuser ceph --setgroup ceph
[tp_osd_tp]
3956201 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 4.72 % ceph-osd -f
--cluster ceph --id 11 --setuser ceph --setgroup ceph
[tp_osd_tp]
3954970 be/4 ceph 361.68 K/s 0.00 B/s 0.00 % 1.98 % ceph-osd -f
--cluster ceph --id 6 --setuser ceph --setgroup ceph
[tp_osd_tp]
3956653 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 1.26 % ceph-osd -f
--cluster ceph --id 2 --setuser ceph --setgroup ceph
[tp_osd_tp]
3957297 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 0.60 % ceph-osd -f
--cluster ceph --id 7 --setuser ceph --setgroup ceph
[tp_osd_tp]
3956648 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 0.14 % ceph-osd -f
--cluster ceph --id 2 --setuser ceph --setgroup ceph
[tp_osd_tp]
Only OSD1 is doing IO, the rest are doing nothing in comparison, so this must
be the scrubbing, but why is it not running as idle?
There are idle threads though, they just don't seem to do anything
iotop --batch --iter 1 | grep ceph | grep idle
3954974 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 6 --setuser ceph --setgroup ceph
[tp_osd_disk]
3956209 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 11 --setuser ceph --setgroup ceph
[tp_osd_disk]
3956655 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 2 --setuser ceph --setgroup ceph
[tp_osd_disk]
3957253 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 8 --setuser ceph --setgroup ceph
[tp_osd_disk]
3957306 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 7 --setuser ceph --setgroup ceph
[tp_osd_disk]
3957350 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 10 --setuser ceph --setgroup ceph
[tp_osd_disk]
3958464 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 1 --setuser ceph --setgroup ceph
[tp_osd_disk]
3958505 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 4 --setuser ceph --setgroup ceph
[tp_osd_disk]
3958832 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 3 --setuser ceph --setgroup ceph
[tp_osd_disk]
3960488 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 5 --setuser ceph --setgroup ceph
[tp_osd_disk]
3960851 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 9 --setuser ceph --setgroup ceph
[tp_osd_disk]
3959437 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f
--cluster ceph --id 0 --setuser ceph --setgroup ceph
[tp_osd_disk]
ceph daemon osd.1 config get osd_disk_thread_ioprio_class
{
"osd_disk_thread_ioprio_class": "idle"
}
ceph daemon osd.1 config get osd_disk_thread_ioprio_priority
{
"osd_disk_thread_ioprio_priority": "7"
}
cat /sys/block/sd*/queue/scheduler
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
noop deadline [cfq]
Any Ideas or have I misunderstood something?
Nick
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com