Hi Anthony,

I dont see any error from dmesg log and smart of those OSDs

Regards

On Fri, 3 Oct 2025 at 07:08 Anthony D'Atri <[email protected]> wrote:

> Check dmesg / var/log/messages on the hosts with that PG’s OSDs for
> errors, and run a `smartctl -a` on each of those OSDs’ drives.
>
> > On Oct 2, 2025, at 7:55 PM, Sa Pham <[email protected]> wrote:
> >
> > Hi Joshua,
> >
> > No. OSD is still responding.
> >
> > # ceph tell osd.130 version
> > {
> >    "version": "18.2.7-0-g6b0e988052e",
> >    "release": "reef",
> >    "release_type": "stable"
> > }
> >
> > But the primary OSD which includes slow pg (18.773) will respond slower.
> >
> >
> > detailed as below
> >
> >
> > # ceph pg dump_stuck degraded
> > PG_STAT  STATE                                            UP
> > UP_PRIMARY  ACTING     ACTING_PRIMARY
> > 18.773   active+undersized+degraded+remapped+backfilling  [302,150,138]
> >    302  [130,101]             130
> >
> >
> > # time ceph tell osd.130 version
> > {
> >    "version": "18.2.7-0-g6b0e988052e",
> >    "release": "reef",
> >    "release_type": "stable"
> > }
> >
> > real    0m2.113s
> > user    0m0.148s
> > sys     0m0.036s
> > # time ceph tell osd.101 version
> > {
> >    "version": "18.2.7-0-g6b0e988052e",
> >    "release": "reef",
> >    "release_type": "stable"
> > }
> >
> > real    0m0.192s
> > user    0m0.152s
> > sys     0m0.037s
> >
> >
> > I don't know why.
> >
> >
> > Regards,
> >
> >
> >
> > On Fri, Oct 3, 2025 at 4:21 AM Joshua Blanch <[email protected]>
> > wrote:
> >
> >> could it be an OSD not responding?
> >>
> >> I would usually do
> >>
> >> ceph tell osd.* version
> >>
> >> To test if you can connect with osds
> >>
> >>
> >> On Thu, Oct 2, 2025 at 12:09 PM Sa Pham <[email protected]> wrote:
> >>
> >>> *Hello everyone,*
> >>>
> >>> I’m running a Ceph cluster used as an RGW backend, and I’m facing an
> issue
> >>> with one particular placement group (PG).
> >>>
> >>>
> >>>   -
> >>>
> >>>   Accessing objects from this PG is *extremely slow*.
> >>>   -
> >>>
> >>>   Even running ceph pg <pg_id> takes a very long time.
> >>>   -
> >>>
> >>>   The PG is currently *stuck in a degraded state*, so I’m unable to
> move
> >>>   it to other OSDs.
> >>>
> >>>
> >>> Current ceph version is reef 18.2.7.
> >>>
> >>> Has anyone encountered a similar issue before or have any suggestions
> on
> >>> how to troubleshoot and resolve it?
> >>>
> >>>
> >>> Thanks in advance!
> >>> _______________________________________________
> >>> ceph-users mailing list -- [email protected]
> >>> To unsubscribe send an email to [email protected]
> >>>
> >>
> >
> > --
> > Sa Pham Dang
> > Skype: great_bn
> > Phone/Telegram: 0986.849.582
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
>
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to