Thanks Matt! I didn't know about nobackfill and norebalance! That could be a 
good stop gap, as long as there's no issue having it set for weeks. We estimate 
our legacy bluestore cleanup to take about 3-4 weeks. 

You are correct, I don't want to cancel it we just need to catch up on other 
maintenance items. First and foremost is our MON db growing like crazy. If this 
doesn't make them compact, I'll send out another email. But I'm pretty sure 
docs say that it won't compact while PGs are not active+clean. 

When the PG drop happens we do see PG status including the term premerge so you 
are probably right about this being caused by a PG merge. 

Thank you,
Ray 
 

-----Original Message-----
From: Matt Vandermeulen <stor...@reenigne.net> 
Sent: Tuesday, April 12, 2022 2:39 PM
To: Ray Cunningham <ray.cunning...@keepertech.com>
Cc: ceph-users@ceph.io
Subject: Re: [ceph-users] Stop Rebalancing

It sounds like this is from a PG merge, so I'm going to _guess_ that you don't 
want to straight up cancel the current backfill and instead pause it to catch 
your breath.

You can set `nobackfill` and/or `norebalance` which should pause the backfill.  
Alternatively, use `ceph config set osd.* osd_max_backfills 0` to stop all OSDs 
from allowing backfill to continue.  You could use this to throttle it on an 
OSD cadence, though that's a bit messy.  
Consider the recovery sleep options for that, too.

However, if you want to fully cancel the rebalance, you might want to set the 
PG count back to where you were (if that's what you want), and unless you had a 
bunch of upmaps already, your cluster should be mostly balanced, minus the data 
that has already PG-merged.

I don't think you can do something like use `pgremapper cancel-backfill --yes` 
(see Github) for this because of the PG merge (though maybe you can, I haven't 
tried it), which will add upmaps for ongoing remapped PGs to stop them from 
happening.

Others can chime in with other options, I'm always interested in new ways to 
reign in lots of backfill.


On 2022-04-12 16:03, Ray Cunningham wrote:
> Hi Everyone,
> 
> We just upgraded our 640 OSD cluster to Ceph 16.2.7 and the resulting
> rebalancing of misplaced objects is overwhelming the cluster and
> impacting MON DB compaction, deep scrub repairs and us upgrading
> legacy bluestore OSDs. We have to pause the rebalancing if misplaced
> objects or we're going to fall over.
> 
> Autoscaler-status tells us that we are reducing our PGs by 700'ish
> which will take us over 100 days to complete at our current recovery
> speed. We disabled autoscaler on our biggest pool, but I'm concerned
> that it's already on the path to the lower PG count and won't stop
> adding to our misplaced count after drop below 5%. What can we do to
> stop the cluster from finding more misplaced objects to rebalance?
> Should we set the PG num manually to what our current count is? Or
> will that cause even more havoc?
> 
> Any other thoughts or ideas? My goals are to stop the rebalancing
> temporarily so we can deep scrub and repair inconsistencies, upgrade
> legacy bluestore OSDs and compact our MON DBs (supposedly MON DBs
> don't compact when you aren't 100% active+clean).
> 
> Thank you,
> Ray
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to