And I confirm that a repair is not useful. As as far I can see it simply
"cleans" the error (without modifying the big object) but the error of
course reappears when the deep scrub runs again on that PG
Cheers, Massimo
On Thu, Jan 16, 2020 at 9:35 AM Massimo Sgaravatto <
In my cluster I saw that the problematic objects have been uploaded by a
specific application (onedata), which I think used to upload the files
doing something like:
rados --pool put
Now (since Luminous ?) the default object size is 128MB but if I am not
wrong it was 100GB before.
This would
I just changed my max object size to 256MB and scrubbed and the errors went
away. I’m not sure what can be done to reduce the size of these objects,
though, if it really is a problem. Our cluster has dynamic bucket index
resharding turned on, but that sharding process shouldn’t help it if
I never changed the default value for that attribute
I am missing why I have such big objects around
I am also wondering what a pg repair would do in such case
Il mer 15 gen 2020, 16:18 Liam Monahan ha scritto:
> Thanks for that link.
>
> Do you have a default osd max object size of 128M?
Thanks for that link.
Do you have a default osd max object size of 128M? I’m thinking about doubling
that limit to 256MB on our cluster. Our largest object is only about 10% over
that limit.
> On Jan 15, 2020, at 3:51 AM, Massimo Sgaravatto
> wrote:
>
> I guess this is coming from:
>
>
I guess this is coming from:
https://github.com/ceph/ceph/pull/30783
introduced in Nautilus 14.2.5
On Wed, Jan 15, 2020 at 8:10 AM Massimo Sgaravatto <
massimo.sgarava...@gmail.com> wrote:
> As I wrote here:
>
>
> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2020-January/037909.html
>
>
As I wrote here:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2020-January/037909.html
I saw the same after an update from Luminous to Nautilus 14.2.6
Cheers, Massimo
On Tue, Jan 14, 2020 at 7:45 PM Liam Monahan wrote:
> Hi,
>
> I am getting one inconsistent object on our cluster with
Hi,
I am getting one inconsistent object on our cluster with an inconsistency error
that I haven’t seen before. This started happening during a rolling upgrade of
the cluster from 14.2.3 -> 14.2.6, but I am not sure that’s related.
I was hoping to know what the error means before trying a
On Tue, Sep 25, 2018 at 7:50 PM Sergey Malinin wrote:
>
> # rados list-inconsistent-obj 1.92
> {"epoch":519,"inconsistents":[]}
It's likely the epoch has changed since the last scrub and you'll need
to run another scrub to repopulate this data.
>
> September 25, 2018 4:58 AM, "Brad Hubbard"
# rados list-inconsistent-obj 1.92
{"epoch":519,"inconsistents":[]}
September 25, 2018 4:58 AM, "Brad Hubbard" wrote:
> What does the output of the following command look like?
>
> $ rados list-inconsistent-obj 1.92
___
ceph-users mailing list
And where is the manual for bluestore?
-Original Message-
From: mj [mailto:li...@merit.unu.edu]
Sent: dinsdag 25 september 2018 9:56
To: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] PG inconsistent, "pg repair" not working
Hi,
I was able to solve a similar is
Hi,
I was able to solve a similar issue on our cluster using this blog:
https://ceph.com/geen-categorie/ceph-manually-repair-object/
It does help if you are running a 3/2 config.
Perhaps it helps you as well.
MJ
On 09/25/2018 02:37 AM, Sergey Malinin wrote:
Hello,
During normal operation
Hello,
During normal operation our cluster suddenly thrown an error and since then we
have had 1 inconsistent PG, and one of clients sharing cephfs mount has started
to occasionally log "ceph: Failed to find inode X".
"ceph pg repair" deep scrubs the PG and fails with the same error in log.
Can
As a part of the repair operation it runs a deep-scrub on the PG. If it
showed active+clean after the repair and deep-scrub finished, then the next
run of a scrub on the PG shouldn't change the PG status at all.
On Wed, Jun 6, 2018 at 8:57 PM Adrian wrote:
> Update to this.
>
> The affected pg
Update to this.
The affected pg didn't seem inconsistent:
[root@admin-ceph1-qh2 ~]# ceph health detail
HEALTH_ERR 1 scrub errors; Possible data damage: 1 pg inconsistent
OSD_SCRUB_ERRORS 1 scrub errors
PG_DAMAGED Possible data damage: 1 pg inconsistent
pg 6.20 is active+clean+inconsistent,
Hi Cephers,
We recently upgraded one of our clusters from hammer to jewel and then to
luminous (12.2.5, 5 mons/mgr, 21 storage nodes * 9 osd's). After some
deep-scubs we have an inconsistent pg with a log message we've not seen
before:
HEALTH_ERR 1 scrub errors; Possible data damage: 1 pg
Hi Brad
Thank you very much for your attention.
On 07.03.2018 23:46, Brad Hubbard wrote:
On Thu, Mar 8, 2018 at 1:22 AM, Harald Staub wrote:
"ceph pg repair" leads to:
5.7bd repair 2 errors, 0 fixed
Only an empty list from:
rados list-inconsistent-obj 5.7bd
On Thu, Mar 8, 2018 at 1:22 AM, Harald Staub wrote:
> "ceph pg repair" leads to:
> 5.7bd repair 2 errors, 0 fixed
>
> Only an empty list from:
> rados list-inconsistent-obj 5.7bd --format=json-pretty
>
> Inspired by http://tracker.ceph.com/issues/12577 , I tried again with
"ceph pg repair" leads to:
5.7bd repair 2 errors, 0 fixed
Only an empty list from:
rados list-inconsistent-obj 5.7bd --format=json-pretty
Inspired by http://tracker.ceph.com/issues/12577 , I tried again with
more verbose logging and searched the osd logs e.g. for "!=",
"mismatch", could not
I found it is similar to bug: http://tracker.ceph.com/issues/21388.
And fix it by rados command.
The pg inconsistent info is like following,wish it could be fixed in the future.
root@n10-075-019:/var/lib/ceph/osd/ceph-27/current/1.fcd_head# rados
list-inconsistent-obj 1.fcd --format=json-pretty
Hi, list,
We ran into pg deep scrub error. And we tried to repair it by `ceph pg
repair pgid`. But it didn't work. We also verified object files, and
found both 3 replicas were zero size. What's the problem, whether it
is a bug? And how to fix the inconsistent? I haven't restarted the
osds so
8ae100 1 bdev(0x7f26de472e00
/var/lib/ceph/osd/ceph-0/block) close
2017-08-09 11:41:25.944548 7f26db8ae100 1 bdev(0x7f26de2b3a00
/var/lib/ceph/osd/ceph-0/block) close
-Original Message-
From: Sage Weil [mailto:s...@newdream.net]
Sent: woensdag 9 augustus 2017 4:44
To: Brad
_impl.cc:343] Shutdown
complete
2017-08-09 11:41:25.686088 7f26db8ae100 1 bluefs umount
2017-08-09 11:41:25.705389 7f26db8ae100 1 bdev(0x7f26de472e00
/var/lib/ceph/osd/ceph-0/block) close
2017-08-09 11:41:25.944548 7f26db8ae100 1 bdev(0x7f26de2b3a00
/var/lib/ceph/osd/ceph-0/block) close
utex(0x7f49fb4d2038, FUTEX_WAKE_PRIVATE, 1
> > 23552 16:26:31.336801 madvise(0x7f4a0cafa000, 2555904, MADV_DONTNEED
> > 23553 16:26:31.336915 <... futex resumed> ) = 0 <0.000113>
> > 23552 16:26:31.336959 <... madvise resumed> ) = 0 <0.000148> 23553
> >
52 16:26:31.338270 madvise(0x7f4a01ae4000, 16384, MADV_DONTNEED) = 0
> <0.20> 23552 16:26:31.338320 madvise(0x7f4a018cc000, 49152, MADV_DONTNEED)
> = 0 <0.14> 23552 16:26:31.338561 madvise(0x7f4a0770a000, 24576,
> MADV_DONTNEED) = 0 <0.15> 23552 16:26:31.339161 madv
:31.339235 madvise(0x7f4a02102000, 32768, MADV_DONTNEED) = 0
<0.14>
23552 16:26:31.339331 madvise(0x7f4a01df8000, 16384, MADV_DONTNEED) = 0
<0.19>
23552 16:26:31.339372 madvise(0x7f4a01df8000, 32768, MADV_DONTNEED) = 0
<0.13>
-Original Message-
From: Brad Hubbard
On Sat, Aug 5, 2017 at 1:21 AM, Marc Roos wrote:
>
> I have got a placement group inconsistency, and saw some manual where
> you can export and import this on another osd. But I am getting an
> export error on every osd.
>
> What does this export_files error -5
:52
To: Marc Roos; ceph-users
Subject: Re: [ceph-users] Pg inconsistent / export_files error -5
It _should_ be enough. What happened in your cluster recently? Power
Outage, OSD failures, upgrade, added new hardware, any changes at all.
What is your Ceph version?
On Fri, Aug 4, 2017 at 11:22 AM
It _should_ be enough. What happened in your cluster recently? Power
Outage, OSD failures, upgrade, added new hardware, any changes at all. What
is your Ceph version?
On Fri, Aug 4, 2017 at 11:22 AM Marc Roos wrote:
>
> I have got a placement group inconsistency, and
I have got a placement group inconsistency, and saw some manual where
you can export and import this on another osd. But I am getting an
export error on every osd.
What does this export_files error -5 actually mean? I thought 3 copies
should be enough to secure your data.
> PG_DAMAGED
Hello,
log [INF] : 3.136 repair ok, 0 fixed
Thank you Greg, I did like that, it worked well.
Laurent
Le 25/11/2013 19:10, Gregory Farnum a écrit :
On Mon, Nov 25, 2013 at 8:10 AM, Laurent Barbe laur...@ksperis.com wrote:
Hello,
Since yesterday, scrub has detected an inconsistent pg :( :
Hello,
Since yesterday, scrub has detected an inconsistent pg :( :
# ceph health detail(ceph version 0.61.9)
HEALTH_ERR 1 pgs inconsistent; 9 scrub errors
pg 3.136 is active+clean+inconsistent, acting [9,1]
9 scrub errors
# ceph pg map 3.136
osdmap e4363 pg 3.136 (3.136) - up [9,1] acting
On Mon, Nov 25, 2013 at 8:10 AM, Laurent Barbe laur...@ksperis.com wrote:
Hello,
Since yesterday, scrub has detected an inconsistent pg :( :
# ceph health detail(ceph version 0.61.9)
HEALTH_ERR 1 pgs inconsistent; 9 scrub errors
pg 3.136 is active+clean+inconsistent, acting [9,1]
9
33 matches
Mail list logo