I created Issue #6278 (http://tracker.ceph.com/issues/6278) to track
this issue.
Thanks,
Mike Dawson
On 8/30/2013 1:52 PM, Andrey Korolyov wrote:
On Fri, Aug 30, 2013 at 9:44 PM, Mike Dawson mike.daw...@cloudapt.com wrote:
Andrey,
I use all the defaults:
# ceph --admin-daemon
We've been struggling with an issue of spikes of high i/o latency with
qemu/rbd guests. As we've been chasing this bug, we've greatly improved
the methods we use to monitor our infrastructure.
It appears that our RBD performance chokes in two situations:
- Deep-Scrub
- Backfill/recovery
In
You may want to reduce scrubbing pgs per osd to 1 using config option
and check the results.
On Fri, Aug 30, 2013 at 8:03 PM, Mike Dawson mike.daw...@cloudapt.com wrote:
We've been struggling with an issue of spikes of high i/o latency with
qemu/rbd guests. As we've been chasing this bug, we've
Andrey,
I use all the defaults:
# ceph --admin-daemon /var/run/ceph/ceph-osd.1.asok config show | grep scrub
osd_scrub_thread_timeout: 60,
osd_scrub_finalize_thread_timeout: 600,
osd_max_scrubs: 1,
osd_scrub_load_threshold: 0.5,
osd_scrub_min_interval: 86400,
osd_scrub_max_interval:
On Fri, Aug 30, 2013 at 9:44 PM, Mike Dawson mike.daw...@cloudapt.com wrote:
Andrey,
I use all the defaults:
# ceph --admin-daemon /var/run/ceph/ceph-osd.1.asok config show | grep scrub
osd_scrub_thread_timeout: 60,
osd_scrub_finalize_thread_timeout: 600,
osd_max_scrubs: 1,
This