You can try adding ---------------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1
On Tue, Oct 22, 2019 at 8:36 PM David Turner <[email protected]> wrote: > > Most times you are better served with simpler settings like > osd_recovery_sleep, which has 3 variants if you have multiple types of OSDs > in your cluster (osd_recovery_sleep_hdd, osd_recovery_sleep_sdd, > osd_recovery_sleep_hybrid). Using those you can tweak a specific type of OSD > that might be having problems during recovery/backfill while allowing the > others to continue to backfill at regular speeds. > > Additionally you mentioned reweighting OSDs, but it sounded like you do this > manually. The balancer module, especially in upmap mode, can be configured > quite well to minimize client IO impact while balancing. You can specify > times of day that it can move data (only in UTC, it ignores local timezones), > a threshold of misplaced data that it will stop moving PGs at, the increment > size it will change weights with per operation, how many weights it will > adjust with each pass, etc. > > On Tue, Oct 22, 2019, 6:07 PM Mark Kirkwood <[email protected]> > wrote: >> >> Thanks - that's a good suggestion! >> >> However I'd still like to know the answers to my 2 questions. >> >> regards >> >> Mark >> >> On 22/10/19 11:22 pm, Paul Emmerich wrote: >> > getting rid of filestore solves most latency spike issues during >> > recovery because they are often caused by random XFS hangs (splitting >> > dirs or just xfs having a bad day) >> > >> > >> > Paul >> > >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
