[ceph-users] Re: Next quincy point release 17.2.7

2023-10-04 Thread Tobias Urdin
Hello Yuri, On the RGW side I would very much like to get this [1] patch in that release that is already merged in reef [2] and pacific [3]. Perhaps Casey can approve and merge that so you can bring it into your testing. Thanks! [1] https://github.com/ceph/ceph/pull/53414 [2]

[ceph-users] Re: Ceph 16.2.x excessive logging, how to reduce?

2023-10-04 Thread Zakhar Kirpichenko
Thank you for your response, Igor. Currently debug_rocksdb is set to 4/5: # ceph config get osd debug_rocksdb 4/5 This setting seems to be default. Is my understanding correct that you're suggesting setting it to 3/5 or even 0/5? Would setting it to 0/5 have any negative effects on the cluster?

[ceph-users] Re: outdated mds slow requests

2023-10-04 Thread Ben
Hi Eugen, warnings continue to spam cluster log.Actually for the whole picture of the issue please see: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/VDL56J75FG5LO4ZECIWWGGBW4ULPZUIP/ I was thinking about the following options: 1, restart problematic nodes: 24,32,34,36: need to

[ceph-users] Re: Next quincy point release 17.2.7

2023-10-04 Thread Satoru Takeuchi
Hi Yuri (resend it because I forgot to add ccs to MLs.) 2023年10月5日(木) 5:57 Yuri Weinstein : Hello We are getting very close to the next Quincy point release 17.2.7 Here is the list of must-have PRs https://pad.ceph.com/p/quincy_17.2.7_prs We will start the release testing/review/approval

[ceph-users] Re: rgw: disallowing bucket creation for specific users?

2023-10-04 Thread Matthias Ferdinand
On Tue, Oct 03, 2023 at 06:10:17PM +0200, Matthias Ferdinand wrote: > On Sun, Oct 01, 2023 at 12:00:58PM +0200, Peter Goron wrote: > > Hi Matthias, > > > > One possible way to achieve your need is to set a quota on number of > > buckets at user level (see > >

[ceph-users] Next quincy point release 17.2.7

2023-10-04 Thread Yuri Weinstein
Hello We are getting very close to the next Quincy point release 17.2.7 Here is the list of must-have PRs https://pad.ceph.com/p/quincy_17.2.7_prs We will start the release testing/review/approval process as soon as all PRs from this list are merged. If you see something missing please speak up

[ceph-users] Question about RGW S3 Select

2023-10-04 Thread Dave S
Hi Everyone, I've been trying to get S3 Select working on our system and whenever I send a query I get the following in the Payload (Result 200 from RGW): # aws --endpoint-url http://cephtest1 s3api select-object-content --bucket test1 --expression-type SQL --input-serialization '{"CSV":

[ceph-users] Re: Ceph 16.2.x excessive logging, how to reduce?

2023-10-04 Thread Igor Fedotov
Hi Zakhar, do reduce rocksdb logging verbosity you might want to set debug_rocksdb to 3 (or 0). I presume it produces a  significant part of the logging traffic. Thanks, Igor On 04/10/2023 20:51, Zakhar Kirpichenko wrote: Any input from anyone, please? On Tue, 19 Sept 2023 at 09:01,

[ceph-users] Re: Ceph 16.2.x excessive logging, how to reduce?

2023-10-04 Thread Zakhar Kirpichenko
Any input from anyone, please? On Tue, 19 Sept 2023 at 09:01, Zakhar Kirpichenko wrote: > Hi, > > Our Ceph 16.2.x cluster managed by cephadm is logging a lot of very > detailed messages, Ceph logs alone on hosts with monitors and several OSDs > has already eaten through 50% of the endurance of

[ceph-users] Manual resharding with multisite

2023-10-04 Thread Yixin Jin
Hi folks, I am aware that dynamic resharding isn't supported before Reef with multisite. However, does manual resharding work? It doesn't seem to be so, either. First of all, running "bucket reshard" has to be in the master zone. But if the objects of that bucket isn't in the master zone,

[ceph-users] Re: rgw: disallowing bucket creation for specific users?

2023-10-04 Thread Matthias Ferdinand
> Tried a negative number ("--max-buckets=-1"), but that had no effect at > all (not even an error message). must have mistyped the command; trying again with "-max-buckets=-1", it shows the wanted effect: user cannot create any bucket. So, an effective and elegant method indeed :-) Matthias

[ceph-users] Re: rgw: disallowing bucket creation for specific users?

2023-10-04 Thread Matthias Ferdinand
On Tue, Oct 03, 2023 at 06:10:17PM +0200, Matthias Ferdinand wrote: > On Sun, Oct 01, 2023 at 12:00:58PM +0200, Peter Goron wrote: > > Hi Matthias, > > > > One possible way to achieve your need is to set a quota on number of > > buckets at user level (see > >

[ceph-users] Calling all Ceph users and developers! Submit a topic for the next User + Dev Meeting!

2023-10-04 Thread Laura Flores
Hi Ceph users and developers, We are gearing up for the next User + Developer Monthly Meeting, happening October 19th at 10am EST. If you are interested in being a guest speaker, you are invited to submit a focus topic to this Google form:

[ceph-users] Re: snap_schedule works after 1 hour of scheduling

2023-10-04 Thread Milind Changire
On Wed, Oct 4, 2023 at 7:19 PM Kushagr Gupta wrote: > > Hi Milind, > > Thank you for your swift response. > > >>How many hours did you wait after the "start time" and decide to restart > >>mgr ? > We waited for ~3 days before restarting the mgr-service. The only thing I can think of is a stale

[ceph-users] Re: Autoscaler problems in pacific

2023-10-04 Thread Boris Behrens
Also found what the 2nd problem was: When there are pools using the default replicated_ruleset while there are multiple rulesets with differenct device classes, the autoscaler does not produce any output. Should I open a bug for that? Am Mi., 4. Okt. 2023 um 14:36 Uhr schrieb Boris Behrens : >

[ceph-users] Re: Remove empty orphaned PGs not mapped to a pool

2023-10-04 Thread Malte Stroem
Hello Eugen, yes, we followed the documentation and everything worked fine. The cache is gone. Removing the pool worked well. Everything is clean. The PGs are empty active+clean. Possible solutions: 1. ceph pg {pg-id} mark_unfound_lost delete I do not think this is the right way since it

[ceph-users] Re: snap_schedule works after 1 hour of scheduling

2023-10-04 Thread Kushagr Gupta
Hi Milind, Thank you for your swift response. >>How many hours did you wait after the "start time" and decide to restart mgr ? We waited for ~3 days before restarting the mgr-service. There was one more instance where we waited for 2 hours and then re-started and in the third hour the schedule

[ceph-users] Issue with radosgw-admin reshard when bucket belongs to user with tenant on ceph quincy (17.2.6)

2023-10-04 Thread christoph . weber+cephmailinglist
Hi everybody, I tried to reshard a bucket belonging to the tenant "test-tenant", but got an "No such file or directory" error. $ radosgw-admin reshard add --bucket test-tenant/test-bucket --num-shards 40 $ radosgw-admin reshard process 2023-10-04T12:12:52.470+0200 7f654237afc0 0

[ceph-users] snap_schedule works after 1 hour of scheduling

2023-10-04 Thread Kushagr Gupta
Hi Team,Milind *Ceph-version:* Quincy, Reef *OS:* Almalinux 8 *Issue:* snap_schedule works after 1 hour of schedule *Description:* We are currently working in a 3-node ceph cluster. We are currently exploring the scheduled snapshot capability of the ceph-mgr module. To enable/configure

[ceph-users] Re: Autoscaler problems in pacific

2023-10-04 Thread Boris Behrens
Found the bug for the TOO_MANY_PGS: https://tracker.ceph.com/issues/62986 But I am still not sure, why I don't have any output on that one cluster. Am Mi., 4. Okt. 2023 um 14:08 Uhr schrieb Boris Behrens : > Hi, > I've just upgraded to our object storages to the latest pacific version >

[ceph-users] Re: snap_schedule works after 1 hour of scheduling

2023-10-04 Thread Milind Changire
On Wed, Oct 4, 2023 at 3:40 PM Kushagr Gupta wrote: > > Hi Team,Milind > > Ceph-version: Quincy, Reef > OS: Almalinux 8 > > Issue: snap_schedule works after 1 hour of schedule > > Description: > > We are currently working in a 3-node ceph cluster. > We are currently exploring the scheduled

[ceph-users] Autoscaler problems in pacific

2023-10-04 Thread Boris Behrens
Hi, I've just upgraded to our object storages to the latest pacific version (16.2.14) and the autscaler is acting weird. On one cluster it just shows nothing: ~# ceph osd pool autoscale-status ~# On the other clusters it shows this when it is set to warn: ~# ceph health detail ... [WRN]

[ceph-users] Re: RGW multisite - requesting help for fixing error_code: 125

2023-10-04 Thread Eugen Block
Hi, I just did this successfully on a test Reef cluster (no multi-site) $ radosgw-admin object rewrite --bucket=bucket1 --object="myfile.txt" where "--object" is the object name. The epoch and the tag have been updated, so I guess it worked. But I also got a segfault on a Octopus test

[ceph-users] Re: ceph luminous client connect to ceph reef always permission denied

2023-10-04 Thread Eugen Block
Hi, I suspect the auth_allow_insecure_global_id_reclaim config option. If you really need this to work you can set $ ceph config set mon auth_allow_insecure_global_id_reclaim true and the client should be able to connect. You will get a warning though: mon is allowing insecure global_id

[ceph-users] Re: outdated mds slow requests

2023-10-04 Thread Eugen Block
Hi, is this still an issue? If so, I would try to either evict the client via admin socket: ceph tell mds.5 client evict [...] --- Evict client session(s) based on a filter alternatively locally on the MDS: cephadm enter mds. ceph daemon mds. client evict or restart the MDS which

[ceph-users] Re: Balancer blocked as autoscaler not acting on scaling change

2023-10-04 Thread Joachim Kraftmayer - ceph ambassador
Hi, we have often seen strange behavior and also interesting pg targets from pg_autoscaler in the last years. That's why we disable it globally. The commands: ceph osd reweight-by-utilization ceph osd test-reweight-by-utilization are from the time before the upmap balancer was introduced and

[ceph-users] Re: Balancer blocked as autoscaler not acting on scaling change

2023-10-04 Thread Eugen Block
Hi, you could change the target_max_misplaced_ratio to 1, the balancer has a default 5% ratio of misplaced objects, see [1] for more information: ceph config get mgr target_max_misplaced_ratio 0.05 [1] https://docs.ceph.com/en/latest/rados/operations/balancer/#throttling Zitat von

[ceph-users] Re: VM hangs when overwriting a file on erasure coded RBD

2023-10-04 Thread Peter Linder
Hi all, I would like to follow up on this, it turns out that overwriting the file doesn't actually hang, but is just super slow, like several minutes. The process is busy in a syscall reading large amounts of what I'm assuming is filesystem metadata until the operation finally completes.

[ceph-users] Re: Remove empty orphaned PGs not mapped to a pool

2023-10-04 Thread Eugen Block
Hi, just for clarity, you're actually talking about the cache tier as described in the docs [1]? And you followed the steps until 'ceph osd tier remove cold-storage hot-storage' successfully? And the pool has been really deleted successfully ('ceph osd pool ls detail')? [1]

[ceph-users] Re: Slow recovery and inaccurate recovery figures since Quincy upgrade

2023-10-04 Thread Sake
Hi,Please take a look at the following thread: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/PWHG6QJ6N2TJEYD2U4AXJAJ23CRPJG4E/#7ZMBM23GXYFIGY52ZWJDY5NUSYSDSYL6In short, the value for "osd_mclock_cost_per_byte_usec_hdd" isn't correct. With the release of 17.2.7 this option will be

[ceph-users] Re: set proxy for ceph installation

2023-10-04 Thread Eugen Block
Did you apply the changes to the containers.conf file on all hosts? The MGR daemon is issuing the cephadm commands on the remote hosts, so it would need that as well. That setup works for me quite well for years now. What distro is your host running on? We mostly use openSUSE or SLES, but

[ceph-users] Re: Slow recovery and inaccurate recovery figures since Quincy upgrade

2023-10-04 Thread Sridhar Seshasayee
To help complete the recovery, you can temporarily try disabling scrub and deep scrub operations by running: ceph osd set noscrub ceph osd set nodeep-scrub This should help speed up the recovery process. Once the recovery is done, you can unset the above scrub flags and revert the mClock profile

[ceph-users] Re: cephfs health warn

2023-10-04 Thread Ben
Hi Venky, thanks for help on this. Will change to multimds with subtree pinning. For the moment, it needs to get the segments list items go by loop of expiring -> expired -> trimmed. It is observed that each problematic mds has a few expiring segment stuck in the road of trimming. the segment