Hey,
I'm trying to find the last two objects on an old data pool for a cephfs
cluster that seemingly aren't found via find && getfattr.
More generically though, is there a low level tool or (C++ or Python)
library for reading and analyzing cephfs metadata?
Daniel
Is the balancer activated? (upmap?)
>
> Once the situation stabilizes, it becomes interesting to think about the
> number of pg/osd =>
> https://docs.ceph.com/en/latest/rados/operations/placement-groups/#managing-pools-that-are-flagged-with-bulk
>
>
> Le mer. 27
Hey,
I'm running ceph version 18.2.1 (reef) but this problem must have existed a
long time before reef.
The documentation says the autoscaler will target 100 pgs per OSD but I'm
only seeing ~10. My erasure encoding is a stripe of 6 data 3 parity.
Could that be the reason? PGs numbers for that EC
ues/57152
>
> Please follow there for updates.
>
> On Mon, Aug 15, 2022 at 4:12 PM Daniel Williams
> wrote:
> >
> > ceph-post-file: a9802e30-0096-410e-b5c0-f2e6d83acfd6
> >
> > On Tue, Aug 16, 2022 at 3:13 AM Patrick Donnelly
> wrote:
> >>
&g
ceph-post-file: a9802e30-0096-410e-b5c0-f2e6d83acfd6
On Tue, Aug 16, 2022 at 3:13 AM Patrick Donnelly
wrote:
> On Mon, Aug 15, 2022 at 11:39 AM Daniel Williams
> wrote:
> >
> > Using ubuntu with apt repository from ceph.
> >
> > Ok that helped me figure out that
n Mon, Aug 15, 2022 at 10:38 AM Daniel Williams
> wrote:
> >
> > My managers are crashing reading the sqlite database for deviceheatlth:
> > .mgr:devicehealth/main.db-journal
> > debug -2> 2022-08-15T11:14:09.184+ 7fa5721b7700 5 cephsqlite:
>
My managers are crashing reading the sqlite database for deviceheatlth:
.mgr:devicehealth/main.db-journal
debug -2> 2022-08-15T11:14:09.184+ 7fa5721b7700 5 cephsqlite:
Read: (client.53284882) [.mgr:devicehealth/main.db-journal] 0x5601da0c0008
4129788~65536
debug -1>
I recently moved to quincy and cephadm.
I noticed that when I moved some drives from one machine to another they at
some point got marked as weight 0 in the crushmap.
The first time that was fine, I just fixed it and figured it was something
I did wrong with moving the drives.
The second time it
Do you think maybe you should issue an immediate change/patch/update to
quincy to change the default to wpq? Given the cluster ending nature of the
problem?
On Wed, Jul 20, 2022 at 4:01 AM Sridhar Seshasayee
wrote:
> Hi Daniel,
>
>
> And further to my theory about the spin lock or similar,
s a
median CPU usage of some multiple of 100% (eg. a multiple of a machine
core/thread usage per osd).
On Tue, Jul 19, 2022 at 4:18 PM Daniel Williams wrote:
> Also never had problems with backfill / rebalance / recovery but now seen
> runaway CPU usage even with very conservative recover
Also never had problems with backfill / rebalance / recovery but now seen
runaway CPU usage even with very conservative recovery settings after
upgrading to quincy from pacific.
osd_recovery_sleep_hdd = 0.1
osd_max_backfills = 1
osd_recovery_max_active = 1
osd_recovery_delay_start = 600
Tried:
Does a stripe aware file copier exist for cephfs to parallelize the copying
of one larger file?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
12 matches
Mail list logo