[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wyll Ingersoll
We did this but oddly enough it is showing the movement of PGS away​ from the 
new, underutilized OSDs instead of TO them as we would expect.

From: Wesley Dillingham 
Sent: Tuesday, August 23, 2022 2:13 PM
To: Wyll Ingersoll 
Cc: ceph-users@ceph.io 
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

https://docs.ceph.com/en/pacific/rados/operations/upmap/

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 1:45 PM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:
Thank you - we have increased backfill settings, but can you elaborate on 
"injecting upmaps" ?

From: Wesley Dillingham mailto:w...@wesdillingham.com>>
Sent: Tuesday, August 23, 2022 1:44 PM
To: Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>>
Cc: ceph-users@ceph.io<mailto:ceph-users@ceph.io> 
mailto:ceph-users@ceph.io>>
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

In that case I would say your options are to make use of injecting upmaps to 
move data off the full osds or to increase the backfill throttle settings to 
make things move faster.

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 1:28 PM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:
Unfortunately, I cannot. The system in question is in a secure location and I 
don't have direct access to it.  The person on site runs the commands I send 
them and the osd tree is correct as far as we can tell. The new hosts and osds 
are in the right place in the tree and have proper weights.  One small 
difference is that the new osds have a class ("hdd"), whereas MOST of the 
pre-existing osds do not have a class designation, this is a cluster that has 
grown and been upgraded over several releases of ceph. Currently it is running 
pacific 16.2.9.  However, removing the class designation on one of the new osds 
did not make any difference so I dont think that is the issue.

The cluster is slowly recovering, but our new OSDs are very lightly used at 
this point, only a few PGs have been assigned to them, though more than zero 
and the number does appear to be slowly (very slowly) growing so recovery is 
happening but very very slowly.





From: Wesley Dillingham mailto:w...@wesdillingham.com>>
Sent: Tuesday, August 23, 2022 1:18 PM
To: Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>>
Cc: ceph-users@ceph.io<mailto:ceph-users@ceph.io> 
mailto:ceph-users@ceph.io>>
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

Can you please send the output of "ceph osd tree"

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:

We have a large cluster with a many osds that are at their nearfull or full 
ratio limit and are thus having problems rebalancing.
We added 2 more storage nodes, each with 20 additional drives  to give the 
cluster room to rebalance.  However, for the past few days, the new OSDs are 
NOT being used and the cluster remains stuck and is not improving.

The crush map is correct, the new hosts and osds are at the correct location, 
but dont seem to be getting used.

Any idea how we can force the full or backfillfull OSDs to start unloading 
their pgs to the newly added ones?

thanks!
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wesley Dillingham
https://docs.ceph.com/en/pacific/rados/operations/upmap/

Respectfully,

*Wes Dillingham*
w...@wesdillingham.com
LinkedIn <http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 1:45 PM Wyll Ingersoll <
wyllys.ingers...@keepertech.com> wrote:

> Thank you - we have increased backfill settings, but can you elaborate on
> "injecting upmaps" ?
> --
> *From:* Wesley Dillingham 
> *Sent:* Tuesday, August 23, 2022 1:44 PM
> *To:* Wyll Ingersoll 
> *Cc:* ceph-users@ceph.io 
> *Subject:* Re: [ceph-users] Full cluster, new OSDS not being used
>
> In that case I would say your options are to make use of injecting upmaps
> to move data off the full osds or to increase the backfill throttle
> settings to make things move faster.
>
> Respectfully,
>
> *Wes Dillingham*
> w...@wesdillingham.com
> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>
>
> On Tue, Aug 23, 2022 at 1:28 PM Wyll Ingersoll <
> wyllys.ingers...@keepertech.com> wrote:
>
> Unfortunately, I cannot. The system in question is in a secure location
> and I don't have direct access to it.  The person on site runs the commands
> I send them and the osd tree is correct as far as we can tell. The new
> hosts and osds are in the right place in the tree and have proper weights.
> One small difference is that the new osds have a class ("hdd"), whereas
> MOST of the pre-existing osds do not have a class designation, this is a
> cluster that has grown and been upgraded over several releases of ceph.
> Currently it is running pacific 16.2.9.  However, removing the class
> designation on one of the new osds did not make any difference so I dont
> think that is the issue.
>
> The cluster is slowly recovering, but our new OSDs are very lightly used
> at this point, only a few PGs have been assigned to them, though more than
> zero and the number does appear to be slowly (very slowly) growing so
> recovery is happening but very very slowly.
>
>
>
>
> --------------
> *From:* Wesley Dillingham 
> *Sent:* Tuesday, August 23, 2022 1:18 PM
> *To:* Wyll Ingersoll 
> *Cc:* ceph-users@ceph.io 
> *Subject:* Re: [ceph-users] Full cluster, new OSDS not being used
>
> Can you please send the output of "ceph osd tree"
>
> Respectfully,
>
> *Wes Dillingham*
> w...@wesdillingham.com
> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>
>
> On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll <
> wyllys.ingers...@keepertech.com> wrote:
>
>
> We have a large cluster with a many osds that are at their nearfull or
> full ratio limit and are thus having problems rebalancing.
> We added 2 more storage nodes, each with 20 additional drives  to give the
> cluster room to rebalance.  However, for the past few days, the new OSDs
> are NOT being used and the cluster remains stuck and is not improving.
>
> The crush map is correct, the new hosts and osds are at the correct
> location, but dont seem to be getting used.
>
> Any idea how we can force the full or backfillfull OSDs to start unloading
> their pgs to the newly added ones?
>
> thanks!
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wyll Ingersoll
Thank you - we have increased backfill settings, but can you elaborate on 
"injecting upmaps" ?

From: Wesley Dillingham 
Sent: Tuesday, August 23, 2022 1:44 PM
To: Wyll Ingersoll 
Cc: ceph-users@ceph.io 
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

In that case I would say your options are to make use of injecting upmaps to 
move data off the full osds or to increase the backfill throttle settings to 
make things move faster.

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 1:28 PM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:
Unfortunately, I cannot. The system in question is in a secure location and I 
don't have direct access to it.  The person on site runs the commands I send 
them and the osd tree is correct as far as we can tell. The new hosts and osds 
are in the right place in the tree and have proper weights.  One small 
difference is that the new osds have a class ("hdd"), whereas MOST of the 
pre-existing osds do not have a class designation, this is a cluster that has 
grown and been upgraded over several releases of ceph. Currently it is running 
pacific 16.2.9.  However, removing the class designation on one of the new osds 
did not make any difference so I dont think that is the issue.

The cluster is slowly recovering, but our new OSDs are very lightly used at 
this point, only a few PGs have been assigned to them, though more than zero 
and the number does appear to be slowly (very slowly) growing so recovery is 
happening but very very slowly.





From: Wesley Dillingham mailto:w...@wesdillingham.com>>
Sent: Tuesday, August 23, 2022 1:18 PM
To: Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>>
Cc: ceph-users@ceph.io<mailto:ceph-users@ceph.io> 
mailto:ceph-users@ceph.io>>
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

Can you please send the output of "ceph osd tree"

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:

We have a large cluster with a many osds that are at their nearfull or full 
ratio limit and are thus having problems rebalancing.
We added 2 more storage nodes, each with 20 additional drives  to give the 
cluster room to rebalance.  However, for the past few days, the new OSDs are 
NOT being used and the cluster remains stuck and is not improving.

The crush map is correct, the new hosts and osds are at the correct location, 
but dont seem to be getting used.

Any idea how we can force the full or backfillfull OSDs to start unloading 
their pgs to the newly added ones?

thanks!
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wesley Dillingham
In that case I would say your options are to make use of injecting upmaps
to move data off the full osds or to increase the backfill throttle
settings to make things move faster.

Respectfully,

*Wes Dillingham*
w...@wesdillingham.com
LinkedIn <http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 1:28 PM Wyll Ingersoll <
wyllys.ingers...@keepertech.com> wrote:

> Unfortunately, I cannot. The system in question is in a secure location
> and I don't have direct access to it.  The person on site runs the commands
> I send them and the osd tree is correct as far as we can tell. The new
> hosts and osds are in the right place in the tree and have proper weights.
> One small difference is that the new osds have a class ("hdd"), whereas
> MOST of the pre-existing osds do not have a class designation, this is a
> cluster that has grown and been upgraded over several releases of ceph.
> Currently it is running pacific 16.2.9.  However, removing the class
> designation on one of the new osds did not make any difference so I dont
> think that is the issue.
>
> The cluster is slowly recovering, but our new OSDs are very lightly used
> at this point, only a few PGs have been assigned to them, though more than
> zero and the number does appear to be slowly (very slowly) growing so
> recovery is happening but very very slowly.
>
>
>
>
> --
> *From:* Wesley Dillingham 
> *Sent:* Tuesday, August 23, 2022 1:18 PM
> *To:* Wyll Ingersoll 
> *Cc:* ceph-users@ceph.io 
> *Subject:* Re: [ceph-users] Full cluster, new OSDS not being used
>
> Can you please send the output of "ceph osd tree"
>
> Respectfully,
>
> *Wes Dillingham*
> w...@wesdillingham.com
> LinkedIn <http://www.linkedin.com/in/wesleydillingham>
>
>
> On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll <
> wyllys.ingers...@keepertech.com> wrote:
>
>
> We have a large cluster with a many osds that are at their nearfull or
> full ratio limit and are thus having problems rebalancing.
> We added 2 more storage nodes, each with 20 additional drives  to give the
> cluster room to rebalance.  However, for the past few days, the new OSDs
> are NOT being used and the cluster remains stuck and is not improving.
>
> The crush map is correct, the new hosts and osds are at the correct
> location, but dont seem to be getting used.
>
> Any idea how we can force the full or backfillfull OSDs to start unloading
> their pgs to the newly added ones?
>
> thanks!
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wyll Ingersoll
Unfortunately, I cannot. The system in question is in a secure location and I 
don't have direct access to it.  The person on site runs the commands I send 
them and the osd tree is correct as far as we can tell. The new hosts and osds 
are in the right place in the tree and have proper weights.  One small 
difference is that the new osds have a class ("hdd"), whereas MOST of the 
pre-existing osds do not have a class designation, this is a cluster that has 
grown and been upgraded over several releases of ceph. Currently it is running 
pacific 16.2.9.  However, removing the class designation on one of the new osds 
did not make any difference so I dont think that is the issue.

The cluster is slowly recovering, but our new OSDs are very lightly used at 
this point, only a few PGs have been assigned to them, though more than zero 
and the number does appear to be slowly (very slowly) growing so recovery is 
happening but very very slowly.





From: Wesley Dillingham 
Sent: Tuesday, August 23, 2022 1:18 PM
To: Wyll Ingersoll 
Cc: ceph-users@ceph.io 
Subject: Re: [ceph-users] Full cluster, new OSDS not being used

Can you please send the output of "ceph osd tree"

Respectfully,

Wes Dillingham
w...@wesdillingham.com<mailto:w...@wesdillingham.com>
LinkedIn<http://www.linkedin.com/in/wesleydillingham>


On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll 
mailto:wyllys.ingers...@keepertech.com>> wrote:

We have a large cluster with a many osds that are at their nearfull or full 
ratio limit and are thus having problems rebalancing.
We added 2 more storage nodes, each with 20 additional drives  to give the 
cluster room to rebalance.  However, for the past few days, the new OSDs are 
NOT being used and the cluster remains stuck and is not improving.

The crush map is correct, the new hosts and osds are at the correct location, 
but dont seem to be getting used.

Any idea how we can force the full or backfillfull OSDs to start unloading 
their pgs to the newly added ones?

thanks!
___
ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
To unsubscribe send an email to 
ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Full cluster, new OSDS not being used

2022-08-23 Thread Wesley Dillingham
Can you please send the output of "ceph osd tree"

Respectfully,

*Wes Dillingham*
w...@wesdillingham.com
LinkedIn 


On Tue, Aug 23, 2022 at 10:53 AM Wyll Ingersoll <
wyllys.ingers...@keepertech.com> wrote:

>
> We have a large cluster with a many osds that are at their nearfull or
> full ratio limit and are thus having problems rebalancing.
> We added 2 more storage nodes, each with 20 additional drives  to give the
> cluster room to rebalance.  However, for the past few days, the new OSDs
> are NOT being used and the cluster remains stuck and is not improving.
>
> The crush map is correct, the new hosts and osds are at the correct
> location, but dont seem to be getting used.
>
> Any idea how we can force the full or backfillfull OSDs to start unloading
> their pgs to the newly added ones?
>
> thanks!
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io