Hi. Use value "osd_recovery_delay_start" example: [root@ceph08 ceph]# ceph --admin-daemon /var/run/ceph/ceph-osd.94.asok config show | grep osd_recovery_delay_start "osd_recovery_delay_start": "10"
2015-03-03 13:13 GMT+03:00 Andrija Panic <[email protected]>: > HI Guys, > > I yesterday removed 1 OSD from cluster (out of 42 OSDs), and it caused > over 37% od the data to rebalance - let's say this is fine (this is when I > removed it frm Crush Map). > > I'm wondering - I have previously set some throtling mechanism, but during > first 1h of rebalancing, my rate of recovery was going up to 1500 MB/s - > and VMs were unusable completely, and then last 4h of the duration of > recover this recovery rate went down to, say, 100-200 MB.s and during this > VM performance was still pretty impacted, but at least I could work more or > a less > > So my question, is this behaviour expected, is throtling here working as > expected, since first 1h was almoust no throtling applied if I check the > recovery rate 1500MB/s and the impact on Vms. > And last 4h seemed pretty fine (although still lot of impact in general) > > I changed these throtling on the fly with: > > ceph tell osd.* injectargs '--osd_recovery_max_active 1' > ceph tell osd.* injectargs '--osd_recovery_op_priority 1' > ceph tell osd.* injectargs '--osd_max_backfills 1' > > My Jorunals are on SSDs (12 OSD per server, of which 6 journals on one > SSD, 6 journals on another SSD) - I have 3 of these hosts. > > Any thought are welcome. > -- > > Andrija Panić > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- С уважением, Фасихов Ирек Нургаязович Моб.: +79229045757
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
