[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Reed Dier
> Yeah, there seems to be a fear that attempting to repair those will > negatively impact performance even more. I disagree and think we should do > them immediately. There shouldn’t really be too much of a noticeable performance hit. Some good documentation here

[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Mark Nelson
On 2/28/23 13:11, Dave Ingram wrote: On Tue, Feb 28, 2023 at 12:56 PM Reed Dier wrote: I think a few other things that could help would be `ceph osd df tree` which will show the hierarchy across different crush domains. Good idea: https://pastebin.com/y07TKt52 Yeah, it looks like

[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Dave Ingram
On Tue, Feb 28, 2023 at 12:56 PM Reed Dier wrote: > I think a few other things that could help would be `ceph osd df tree` > which will show the hierarchy across different crush domains. > Good idea: https://pastebin.com/y07TKt52 > And if you’re doing something like erasure coded pools, or

[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Dave Ingram
When I suggested this to the senior admin here I was told that was a bad idea because it would negatively impact performance. Is that true? I thought all that would do was accept the information from the other 2 OSDs and the one with the errors would rebuild the record. The underlying disks

[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Reed Dier
I think a few other things that could help would be `ceph osd df tree` which will show the hierarchy across different crush domains. And if you’re doing something like erasure coded pools, or something other than replication 3, maybe `ceph osd crush rule dump` may provide some further context

[ceph-users] Re: Ceph OSD imbalance and performance

2023-02-28 Thread Janne Johansson
Den tis 28 feb. 2023 kl 18:13 skrev Dave Ingram : > There are also several > scrub errors. In short, it's a complete wreck. > > health: HEALTH_ERR > 3 scrub errors > Possible data damage: 3 pgs inconsistent > [root@ceph-admin davei]# ceph health detail > HEALTH_ERR 3