Hi Frédéric,

I tried to repeer and deep scrub, but it's not working.

Have you already checked the logs for osd.302 and /var/log/messages for any
I/O-related issues?

=> I checked , there is no I/O error/issue.

Regards,

On Mon, Oct 6, 2025 at 3:15 PM Frédéric Nass <[email protected]>
wrote:

> Hi Sa,
>
> Regarding the output you provided, it appears that osd.302 is listed as UP
> but not ACTING for PG 18.773:
>
> PG_STAT  STATE
>              UP                     UP_PRIMARY  ACTING
>  ACTING_PRIMARY
> 18.773       active+undersized+degraded+remapped+backfilling
>  [302,150,138]    302                   [130,101]             130
>
> Have you already checked the logs for osd.302 and /var/log/messages for
> any I/O-related issues? Could you also try running 'ceph pg repeer 18.773'?
>
> If this is the only PG for which `osd.302` is not acting and the 'repeer'
> command does not resolve the issue, I would suggest attempting a deep-scrub
> on this PG.
> This might uncover errors that could potentially be fixed, either online
> or offline.
>
> Regards,
> Frédéric
>
> --
> Frédéric Nass
> Ceph Ambassador France | Senior Ceph Engineer @ CLYSO
> Try our Ceph Analyzer -- https://analyzer.clyso.com/
> https://clyso.com | [email protected]
>
>
> Le lun. 6 oct. 2025 à 06:31, Sa Pham <[email protected]> a écrit :
>
>> Hello Eugen,
>>
>>
>> This PG include: 254490 objects, size: 68095493667 bytes
>>
>>
>> Regards,
>>
>> On Fri, Oct 3, 2025 at 9:10 PM Eugen Block <[email protected]> wrote:
>>
>> > Is it possible that this is a huge PG? What size does it have? But it
>> > could also be a faulty disk.
>> >
>> >
>> > Zitat von Sa Pham <[email protected]>:
>> >
>> > > *Hello everyone,*
>> > >
>> > > I’m running a Ceph cluster used as an RGW backend, and I’m facing an
>> > issue
>> > > with one particular placement group (PG).
>> > >
>> > >
>> > >    -
>> > >
>> > >    Accessing objects from this PG is *extremely slow*.
>> > >    -
>> > >
>> > >    Even running ceph pg <pg_id> takes a very long time.
>> > >    -
>> > >
>> > >    The PG is currently *stuck in a degraded state*, so I’m unable to
>> move
>> > >    it to other OSDs.
>> > >
>> > >
>> > > Current ceph version is reef 18.2.7.
>> > >
>> > > Has anyone encountered a similar issue before or have any suggestions
>> on
>> > > how to troubleshoot and resolve it?
>> > >
>> > >
>> > > Thanks in advance!
>> > > _______________________________________________
>> > > ceph-users mailing list -- [email protected]
>> > > To unsubscribe send an email to [email protected]
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list -- [email protected]
>> > To unsubscribe send an email to [email protected]
>> >
>>
>>
>> --
>> Sa Pham Dang
>> Skype: great_bn
>> Phone/Telegram: 0986.849.582
>> _______________________________________________
>> ceph-users mailing list -- [email protected]
>> To unsubscribe send an email to [email protected]
>>
>

-- 
Sa Pham Dang
Skype: great_bn
Phone/Telegram: 0986.849.582
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to