Could be an issue with the primary OSD which is now osd.130. Have you
checked osd.130 for any errors?
Maybe try restarting osd.130 and osd.302 one at a time and maybe 101 as
well, waiting for ~all PGs to become active+clean between all restarts.

Could you please share a ceph status? So we get a better view of the
situation.

Regards,
Frédéric.

--
Frédéric Nass
Ceph Ambassador France | Senior Ceph Engineer @ CLYSO
Try our Ceph Analyzer -- https://analyzer.clyso.com/
https://clyso.com | [email protected]

Le lun. 6 oct. 2025 à 14:19, Sa Pham <[email protected]> a écrit :

> Hi Frédéric,
>
> I tried to repeer and deep scrub, but it's not working.
>
> Have you already checked the logs for osd.302 and /var/log/messages for
> any I/O-related issues?
>
> => I checked , there is no I/O error/issue.
>
> Regards,
>
> On Mon, Oct 6, 2025 at 3:15 PM Frédéric Nass <[email protected]>
> wrote:
>
>> Hi Sa,
>>
>> Regarding the output you provided, it appears that osd.302 is listed as
>> UP but not ACTING for PG 18.773:
>>
>> PG_STAT  STATE
>>              UP                     UP_PRIMARY  ACTING
>>  ACTING_PRIMARY
>> 18.773       active+undersized+degraded+remapped+backfilling
>>  [302,150,138]    302                   [130,101]             130
>>
>> Have you already checked the logs for osd.302 and /var/log/messages for
>> any I/O-related issues? Could you also try running 'ceph pg repeer 18.773'?
>>
>> If this is the only PG for which `osd.302` is not acting and the 'repeer'
>> command does not resolve the issue, I would suggest attempting a deep-scrub
>> on this PG.
>> This might uncover errors that could potentially be fixed, either online
>> or offline.
>>
>> Regards,
>> Frédéric
>>
>> --
>> Frédéric Nass
>> Ceph Ambassador France | Senior Ceph Engineer @ CLYSO
>> Try our Ceph Analyzer -- https://analyzer.clyso.com/
>> https://clyso.com | [email protected]
>>
>>
>> Le lun. 6 oct. 2025 à 06:31, Sa Pham <[email protected]> a écrit :
>>
>>> Hello Eugen,
>>>
>>>
>>> This PG include: 254490 objects, size: 68095493667 bytes
>>>
>>>
>>> Regards,
>>>
>>> On Fri, Oct 3, 2025 at 9:10 PM Eugen Block <[email protected]> wrote:
>>>
>>> > Is it possible that this is a huge PG? What size does it have? But it
>>> > could also be a faulty disk.
>>> >
>>> >
>>> > Zitat von Sa Pham <[email protected]>:
>>> >
>>> > > *Hello everyone,*
>>> > >
>>> > > I’m running a Ceph cluster used as an RGW backend, and I’m facing an
>>> > issue
>>> > > with one particular placement group (PG).
>>> > >
>>> > >
>>> > >    -
>>> > >
>>> > >    Accessing objects from this PG is *extremely slow*.
>>> > >    -
>>> > >
>>> > >    Even running ceph pg <pg_id> takes a very long time.
>>> > >    -
>>> > >
>>> > >    The PG is currently *stuck in a degraded state*, so I’m unable to
>>> move
>>> > >    it to other OSDs.
>>> > >
>>> > >
>>> > > Current ceph version is reef 18.2.7.
>>> > >
>>> > > Has anyone encountered a similar issue before or have any
>>> suggestions on
>>> > > how to troubleshoot and resolve it?
>>> > >
>>> > >
>>> > > Thanks in advance!
>>> > > _______________________________________________
>>> > > ceph-users mailing list -- [email protected]
>>> > > To unsubscribe send an email to [email protected]
>>> >
>>> >
>>> > _______________________________________________
>>> > ceph-users mailing list -- [email protected]
>>> > To unsubscribe send an email to [email protected]
>>> >
>>>
>>>
>>> --
>>> Sa Pham Dang
>>> Skype: great_bn
>>> Phone/Telegram: 0986.849.582
>>> _______________________________________________
>>> ceph-users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>>
>>
>
> --
> Sa Pham Dang
> Skype: great_bn
> Phone/Telegram: 0986.849.582
>
>
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to