Eugen,

I didn't really think my cluster was eating itself, but I also didn't want
to be in denial.

Regarding the autoscaler, I really thought that it only went up - I didn't
expect that it would decrease the number of PGs.  Plus, I thought I had it
turned off.  I see now that it's off globally but enabled for this
particular pool.  Also, I see that the target PG count is lower than the
current.

I guess you learn something new every day.

-Dave

--
Dave Hall
Binghamton University
kdh...@binghamton.edu
607-760-2328 (Cell)
607-777-4641 (Office)


On Mon, Mar 29, 2021 at 7:52 AM Eugen Block <ebl...@nde.ag> wrote:

> Hi,
>
> that sounds like the pg_autoscaler is doing its work. Check with:
>
> ceph osd pool autoscale-status
>
> I don't think ceph is eating itself or that you're losing data. ;-)
>
>
> Zitat von Dave Hall <kdh...@binghamton.edu>:
>
> > Hello,
> >
> > About 3 weeks ago I added a node and increased the number of OSDs in my
> > cluster from 24 to 32, and then marked one old OSD down because it was
> > frequently crashing.  .
> >
> > After adding the new OSDs the PG count jumped fairly dramatically, but
> ever
> > since, amidst a continuous low level of rebalancing, the number of PGs
> has
> > gradually decreased to less by 25% from it's max value.  Although I don't
> > have specific notes, my perception is that the current number of PGs is
> > actually lower than it was before I added OSDs.
> >
> > So what's going on here?  It is possible to imagine that my cluster is
> > slowly eating itself, and that I'm about to lose 200TB of data. It's also
> > possible to imagine that this is all due to the gradual optimization of
> the
> > pools.
> >
> > Note that the primary pool is an EC 8 + 2 containing about 124TB.
> >
> > Thanks.
> >
> > -Dave
> >
> > --
> > Dave Hall
> > Binghamton University
> > kdh...@binghamton.edu
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@ceph.io
> > To unsubscribe send an email to ceph-users-le...@ceph.io
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to