All, In looking at the options for setting the default pg autoscale option, I notice that there is a global option setting and a per-pool option setting. It seems that the options at the pool level are off, warn, and on. The same, I assume for the global setting.
Is there a way to get rid of the per-pool setting and set the pool to honor the global setting? I think I'm looking for 'off, warn, on, or global'. It seems that once the per-pool option is set for all of one's pools, the global value is irrelevant. This also implies that in a circumstance where one would want to temporarily suspend autoscaling it would be required to modify the setting for each pool and then to modify it back afterward. Thoughts? Thanks -Dave -- Dave Hall Binghamton University kdh...@binghamton.edu On Mon, Mar 29, 2021 at 1:44 PM Anthony D'Atri <anthony.da...@gmail.com> wrote: > Yes the PG autoscalar has a way of reducing PG count way too far. There’s > a claim that it’s better in Pacific, but I tend to recommend disabling it > and calculating / setting pg_num manually. > > > On Mar 29, 2021, at 9:06 AM, Dave Hall <kdh...@binghamton.edu> wrote: > > > > Eugen, > > > > I didn't really think my cluster was eating itself, but I also didn't > want > > to be in denial. > > > > Regarding the autoscaler, I really thought that it only went up - I > didn't > > expect that it would decrease the number of PGs. Plus, I thought I had > it > > turned off. I see now that it's off globally but enabled for this > > particular pool. Also, I see that the target PG count is lower than the > > current. > > > > I guess you learn something new every day. > > > > -Dave > > > > -- > > Dave Hall > > Binghamton University > > kdh...@binghamton.edu > > 607-760-2328 (Cell) > > 607-777-4641 (Office) > > > > > > On Mon, Mar 29, 2021 at 7:52 AM Eugen Block <ebl...@nde.ag> wrote: > > > >> Hi, > >> > >> that sounds like the pg_autoscaler is doing its work. Check with: > >> > >> ceph osd pool autoscale-status > >> > >> I don't think ceph is eating itself or that you're losing data. ;-) > >> > >> > >> Zitat von Dave Hall <kdh...@binghamton.edu>: > >> > >>> Hello, > >>> > >>> About 3 weeks ago I added a node and increased the number of OSDs in my > >>> cluster from 24 to 32, and then marked one old OSD down because it was > >>> frequently crashing. . > >>> > >>> After adding the new OSDs the PG count jumped fairly dramatically, but > >> ever > >>> since, amidst a continuous low level of rebalancing, the number of PGs > >> has > >>> gradually decreased to less by 25% from it's max value. Although I > don't > >>> have specific notes, my perception is that the current number of PGs is > >>> actually lower than it was before I added OSDs. > >>> > >>> So what's going on here? It is possible to imagine that my cluster is > >>> slowly eating itself, and that I'm about to lose 200TB of data. It's > also > >>> possible to imagine that this is all due to the gradual optimization of > >> the > >>> pools. > >>> > >>> Note that the primary pool is an EC 8 + 2 containing about 124TB. > >>> > >>> Thanks. > >>> > >>> -Dave > >>> > >>> -- > >>> Dave Hall > >>> Binghamton University > >>> kdh...@binghamton.edu > >>> _______________________________________________ > >>> ceph-users mailing list -- ceph-users@ceph.io > >>> To unsubscribe send an email to ceph-users-le...@ceph.io > >> > >> > >> _______________________________________________ > >> ceph-users mailing list -- ceph-users@ceph.io > >> To unsubscribe send an email to ceph-users-le...@ceph.io > >> > > _______________________________________________ > > ceph-users mailing list -- ceph-users@ceph.io > > To unsubscribe send an email to ceph-users-le...@ceph.io > > _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io