If you want to reweight only once when you have a failed disk that is being 
balanced off of, set the crush weight for that osd to 0.0.  Then when you fully 
remove the disk from the cluster it will not do any additional backfilling.  
Any change to the crush map will likely move data around, even if you're 
removing an already "removed" osd.

________________________________

[cid:imagebf1f6b.JPG@cb3472a2.40a10c92]<https://storagecraft.com>       David 
Turner | Cloud Operations Engineer | StorageCraft Technology 
Corporation<https://storagecraft.com>
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2760 | Mobile: 385.224.2943

________________________________

If you are not the intended recipient of this message or received it 
erroneously, please notify the sender and delete it, together with any 
attachments, and be advised that any dissemination or copying of this message 
is prohibited.

________________________________

________________________________
From: M Ranga Swami Reddy [swamire...@gmail.com]
Sent: Thursday, December 01, 2016 11:45 PM
To: David Turner
Cc: ceph-users
Subject: Re: [ceph-users] node and its OSDs down...

Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which started the 
recovery.
My worries is - why Ceph is doing the recovery, if an OSD is already down and 
no more in the cluster. That means, ceph already maintained down OSDs objects 
copied to another OSDs.. here is the ceph osd tree o/p:
===
227     0.91                            osd.227 down    0
....
250     0.91                            osd.250 down    0
===

So to avoid the recovery/rebalance , can I set the weight of OSD (which was in 
down state). But is this weight setting also lead to rebalance activity.

Thanks
Swami


On Thu, Dec 1, 2016 at 8:07 PM, David Turner 
<david.tur...@storagecraft.com<mailto:david.tur...@storagecraft.com>> wrote:

I assume you also did ceph osd crush remove osd.<id>.  When you removed the osd 
that was down/out and balanced off of, you changed the weight of the host that 
it was on which triggers additional backfilling to balance the crush map.

________________________________

[cid:image0b480d.JPG@7a964f55.48b534e9]<https://storagecraft.com>       David 
Turner | Cloud Operations Engineer | StorageCraft Technology 
Corporation<https://storagecraft.com>
380 Data Drive Suite 300 | Draper | Utah | 84020
Office: 801.871.2760<tel:(801)%20871-2760> | Mobile: 
385.224.2943<tel:(385)%20224-2943>

________________________________

If you are not the intended recipient of this message or received it 
erroneously, please notify the sender and delete it, together with any 
attachments, and be advised that any dissemination or copying of this message 
is prohibited.

________________________________

________________________________
From: ceph-users 
[ceph-users-boun...@lists.ceph.com<mailto:ceph-users-boun...@lists.ceph.com>] 
on behalf of M Ranga Swami Reddy 
[swamire...@gmail.com<mailto:swamire...@gmail.com>]
Sent: Thursday, December 01, 2016 3:03 AM
To: ceph-users
Subject: [ceph-users] node and its OSDs down...

Hello,
One of my ceph node with 20 OSDs down...After a couple of hours, ceph health is 
in OK state.

Now, I tried to remove those OSDs, which were down state from ceph cluster...
using the "ceh osd remove osd.<id>"
then ceph clsuter started rebalancing...which is strange ..because thsoe OSDs 
are down for a long time and health also OK..
my question - why recovery or reblance started when I remove the OSD (which was 
down).

Thanks
Swami

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to