There are two concerns with setting the reweight to 1.0. The first is peering 
and the second is backfilling. Peering is going to block client I/O on the 
affected OSDs, while backfilling will only potentially slow things down.

I don't know what your client I/O looks like, but personally I would probably 
set the norecover and nobackfill flags, slowly increment your reweight value by 
0.01 or whatever you deem to be appropriate for your environment, waiting for 
peering to complete in between each step. Also allow any resulting blocked 
requests to clear up before incrementing your reweight again.

When your reweight is all the way up to 1.0, inject osd_max_backfills to 
whatever you like (or don't if you're happy with it as is) and unset the 
norecover and nobackfill flags to let backfilling begin. If you are unable to 
handle the impact of backfilling with osd_max_backfills set to 1, then you need 
to add some new OSDs to your cluster before doing any of this. They will have 
to backfill too, but at least you'll have more spindles to handle it.


________________________________

[cid:SC_LOGO_VERT_4C_100x72_f823be1a-ae53-43d3-975c-b054a1b22ec3.jpg]


Steve Taylor | Senior Software Engineer | StorageCraft Technology 
Corporation<https://storagecraft.com>
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2799 |


________________________________
If you are not the intended recipient of this message or received it 
erroneously, please notify the sender and delete it, together with any 
attachments, and be advised that any dissemination or copying of this message 
is prohibited.
________________________________


On Mon, 2018-01-29 at 22:43 +0100, David Majchrzak wrote:

And so I totally forgot to add df tree to the mail.
Here's the interesting bit from two first nodes. where osd.11 has weight but is 
reweighted to 0.


root@osd1:~# ceph osd df tree
ID WEIGHT    REWEIGHT SIZE   USE    AVAIL  %USE  VAR  TYPE NAME
-1 181.99997        -   109T 50848G 60878G     0    0 root default
-2  36.39999        - 37242G 16792G 20449G 45.09 0.99     host osd1
 0   3.64000  1.00000  3724G  1730G  1993G 46.48 1.02         osd.0
 1   3.64000  1.00000  3724G  1666G  2057G 44.75 0.98         osd.1
 2   3.64000  1.00000  3724G  1734G  1989G 46.57 1.02         osd.2
 3   3.64000  1.00000  3724G  1387G  2336G 37.25 0.82         osd.3
 4   3.64000  1.00000  3724G  1722G  2002G 46.24 1.01         osd.4
 6   3.64000  1.00000  3724G  1840G  1883G 49.43 1.08         osd.6
 7   3.64000  1.00000  3724G  1651G  2072G 44.34 0.97         osd.7
 8   3.64000  1.00000  3724G  1747G  1976G 46.93 1.03         osd.8
 9   3.64000  1.00000  3724G  1697G  2026G 45.58 1.00         osd.9
 5   3.64000  1.00000  3724G  1614G  2109G 43.34 0.95         osd.5
-3  36.39999        -      0      0      0     0    0     host osd2
12   3.64000  1.00000  3724G  1730G  1993G 46.46 1.02         osd.12
13   3.64000  1.00000  3724G  1745G  1978G 46.88 1.03         osd.13
14   3.64000  1.00000  3724G  1707G  2016G 45.84 1.01         osd.14
15   3.64000  1.00000  3724G  1540G  2184G 41.35 0.91         osd.15
16   3.64000  1.00000  3724G  1484G  2239G 39.86 0.87         osd.16
18   3.64000  1.00000  3724G  1928G  1796G 51.77 1.14         osd.18
20   3.64000  1.00000  3724G  1767G  1956G 47.45 1.04         osd.20
10   3.64000  1.00000  3724G  1797G  1926G 48.27 1.06         osd.10
49   3.64000  1.00000  3724G  1847G  1877G 49.60 1.09         osd.49
11   3.64000        0      0      0      0     0    0         osd.11



29 jan. 2018 kl. 22:40 skrev David Majchrzak 
<da...@visions.se<mailto:da...@visions.se>>:

Hi!

Cluster: 5 HW nodes, 10 HDDs with SSD journals, filestore, 0.94.9 hammer, 
debian wheezy (scheduled to upgrade once this is fixed).

I have a replaced HDD that another admin set to reweight 0 instead of weight 0 
(I can't remember the reason).
What would be the best way to slowly backfill it? Usually I'm using weight and 
slowly growing it to max size.

I guess if I just set reweight to 1.0, it will backfill as fast as I let it, 
that is max 1 backfill / osd but it will probably disrupt client io (this being 
on hammer).

And if I set the weight on it to 0, the node will get less weight, and will 
start moving data around everywhere right?

Can I use reweight the same way as weight here, slowly increasing it up to 1.0 
by increments of say 0.01?

Kind Regards,
David Majchrzak

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to