I have tried ceph pg repair several times. It claims "instructing pg
2.798s0 on osd.41 to repair" but then nothing happens as far as I can tell.
Any way of knowing if it's doing more?

On Sat, May 18, 2019 at 3:33 PM Brett Chancellor <[email protected]>
wrote:

> I would try the ceph pg repair. If you see the pg go into deep scrubbing,
> then back to inconsistent you probably have a bad drive. Find which of the
> drives in the pg are bad (pg query or go to the host and look through
> dmesg). Take that osd offline and mark it out. Once backfill is complete,
> it should clear up.
>
> On Sat, May 18, 2019, 6:05 PM Jorge Garcia <[email protected]> wrote:
>
>> We are testing a ceph cluster mostly using cephfs. We are using an
>> erasure-code pool, and have been loading it up with data. Recently, we got
>> a HEALTH_ERR response when we were querying the ceph status. We stopped all
>> activity to the filesystem, and waited to see if the error would go away.
>> It didn't. Then we tried a couple of suggestions from the internet (ceph pg
>> repair, ceph pg scrub, ceph pg deep-scrub) to no avail. I'm not sure how to
>> find out more information about what the problem is, and how to repair the
>> filesystem to bring it back to normal health. Any suggestions?
>>
>> Current status:
>>
>> # ceph -s
>>
>>   cluster:
>>
>>     id:     28ef32f1-4350-491b-9003-b19b9c3a2076
>>
>>     health: HEALTH_ERR
>>
>>             5 scrub errors
>>
>>             Possible data damage: 1 pg inconsistent
>>
>>
>>
>>   services:
>>
>>     mon: 3 daemons, quorum gi-cba-01,gi-cba-02,gi-cba-03
>>
>>     mgr: gi-cba-01(active), standbys: gi-cba-02, gi-cba-03
>>
>>     mds: backups-1/1/1 up  {0=gi-cbmd=up:active}
>>
>>     osd: 87 osds: 87 up, 87 in
>>
>>
>>
>>   data:
>>
>>     pools:   2 pools, 4096 pgs
>>
>>     objects: 90.98 M objects, 134 TiB
>>
>>     usage:   210 TiB used, 845 TiB / 1.0 PiB avail
>>
>>     pgs:     4088 active+clean
>>
>>              5    active+clean+scrubbing+deep
>>
>>              2    active+clean+scrubbing
>>
>>              1    active+clean+inconsistent
>>
>> # ceph health detail
>>
>> HEALTH_ERR 5 scrub errors; Possible data damage: 1 pg inconsistent
>>
>> OSD_SCRUB_ERRORS 5 scrub errors
>>
>> PG_DAMAGED Possible data damage: 1 pg inconsistent
>>
>>     pg 2.798 is active+clean+inconsistent, acting [41,50,17,2,86,70,61]
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to