Ah sadly those logs don't look like they have enough debugging to be of
much use.

But from what I'm seeing here, I don't think this state should actually
hurt anything. It ought to go away the next time you delete a snapshot
(maybe only if it has data in those PGs? not sure) and otherwise be
ignorable. I've created a ticket at http://tracker.ceph.com/issues/24876
-Greg

On Wed, Jul 11, 2018 at 9:00 AM Flash <flash...@gmail.com> wrote:

> I found only something like this:
> 2018-07-04 19:26:20.209791 7fc8c8ad2700 -1 log_channel(cluster) log [ERR]
> : trim_object: Can not trim
> 11:e4d50fa4:::rbd_data.4d427a238e1f29.0000000000190c9b:6b4 repair needed
> (no obc)
>
> The full logs in attach. Problem with pg 11.127 started  at ~19:26.
> Later  11.9 became to error state too, but I don't know, at what time
> exactly
>
> ср, 11 июл. 2018 г. в 18:20, Gregory Farnum <gfar...@redhat.com>:
>
>> On Wed, Jul 11, 2018 at 8:07 AM Flash <flash...@gmail.com> wrote:
>>
>>> Hi there.
>>>
>>> Yesterday I caught that error:
>>> PG_DAMAGED Possible data damage: 2 pgs snaptrim_error
>>>     pg 11.9 is active+clean+snaptrim_error, acting [196,167,32]
>>>     pg 11.127 is active+clean+snaptrim_error, acting [184,138,1]
>>> May it be because the scrub was done when the snapshots were cleaned up?
>>>
>>
>> Hmm, the only way you can get the snaptrim_error state is if the PG gets
>> an error when it tries to trim a particular snapshotted object. And it
>> doesn't get cleared by scrubbing; only when it starts snaptrimming again.
>>
>> If you have any OSD logs of when this happened, that would be helpful.
>>
>> And, uh, Sage? Do you know what was supposed to happen here? It's a bit
>> odd as a PG state.
>> -Greg
>>
>>
>>>
>>> I tried to restart OSD, then I run deep-scrub and repair, but it didn't
>>> solve the problem.
>>>
>>> In the documentation the page "Repairing PG inconsistencies" is empty  -
>>>  http://docs.ceph.com/docs/mimic/rados/operations/pg-repair/,
>>> so I don't know, what else can I do?
>>>
>>> Cluster info;
>>> vaersion 12.2.5
>>> 25 OSD nodes
>>> 12 OSD per node. The most of them still have filestore as storage
>>> backend.
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to