Hi Frédéric, I tried to repeer and deep scrub, but it's not working.
Have you already checked the logs for osd.302 and /var/log/messages for any I/O-related issues? => I checked , there is no I/O error/issue. Regards, On Mon, Oct 6, 2025 at 3:15 PM Frédéric Nass <[email protected]> wrote: > Hi Sa, > > Regarding the output you provided, it appears that osd.302 is listed as UP > but not ACTING for PG 18.773: > > PG_STAT STATE > UP UP_PRIMARY ACTING > ACTING_PRIMARY > 18.773 active+undersized+degraded+remapped+backfilling > [302,150,138] 302 [130,101] 130 > > Have you already checked the logs for osd.302 and /var/log/messages for > any I/O-related issues? Could you also try running 'ceph pg repeer 18.773'? > > If this is the only PG for which `osd.302` is not acting and the 'repeer' > command does not resolve the issue, I would suggest attempting a deep-scrub > on this PG. > This might uncover errors that could potentially be fixed, either online > or offline. > > Regards, > Frédéric > > -- > Frédéric Nass > Ceph Ambassador France | Senior Ceph Engineer @ CLYSO > Try our Ceph Analyzer -- https://analyzer.clyso.com/ > https://clyso.com | [email protected] > > > Le lun. 6 oct. 2025 à 06:31, Sa Pham <[email protected]> a écrit : > >> Hello Eugen, >> >> >> This PG include: 254490 objects, size: 68095493667 bytes >> >> >> Regards, >> >> On Fri, Oct 3, 2025 at 9:10 PM Eugen Block <[email protected]> wrote: >> >> > Is it possible that this is a huge PG? What size does it have? But it >> > could also be a faulty disk. >> > >> > >> > Zitat von Sa Pham <[email protected]>: >> > >> > > *Hello everyone,* >> > > >> > > I’m running a Ceph cluster used as an RGW backend, and I’m facing an >> > issue >> > > with one particular placement group (PG). >> > > >> > > >> > > - >> > > >> > > Accessing objects from this PG is *extremely slow*. >> > > - >> > > >> > > Even running ceph pg <pg_id> takes a very long time. >> > > - >> > > >> > > The PG is currently *stuck in a degraded state*, so I’m unable to >> move >> > > it to other OSDs. >> > > >> > > >> > > Current ceph version is reef 18.2.7. >> > > >> > > Has anyone encountered a similar issue before or have any suggestions >> on >> > > how to troubleshoot and resolve it? >> > > >> > > >> > > Thanks in advance! >> > > _______________________________________________ >> > > ceph-users mailing list -- [email protected] >> > > To unsubscribe send an email to [email protected] >> > >> > >> > _______________________________________________ >> > ceph-users mailing list -- [email protected] >> > To unsubscribe send an email to [email protected] >> > >> >> >> -- >> Sa Pham Dang >> Skype: great_bn >> Phone/Telegram: 0986.849.582 >> _______________________________________________ >> ceph-users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] >> > -- Sa Pham Dang Skype: great_bn Phone/Telegram: 0986.849.582 _______________________________________________ ceph-users mailing list -- [email protected] To unsubscribe send an email to [email protected]
