Take a look at https://github.com/TheJJ/ceph-balancer
We switched to it after lot of attempts to make internal balancer work
as expected and now we have ~even OSD utilization across cluster:
# ./placementoptimizer.py -v balance --ensure-optimal-moves
--ensure-variance-decrease
[2023-08-03 23
On 03-Aug-23 12:11 PM, Eugen Block wrote:
ceph balancer status
I changed the PGs and it started rebalancing (and turned autoscaler off)
, so now it will not report status:
It reports: "optimize_result": "Too many objects (0.088184 > 0.05)
are misplaced; try again later"
Lets wait a fe
Turn off the autoscaler and increase pg_num to 512 or so (power of 2).
The recommendation is to have between 100 and 150 PGs per OSD (incl.
replicas). And then let the balancer handle the rest. What is the
current balancer status (ceph balancer status)?
Zitat von Spiros Papageorgiou :
Hi
I set the target_size_ratio of pools by mistake as multiple pools sharing
the same raw capacity. After I adjust it, a large number of pgs are in the
backfill state, but the usage rate of osds is still growing, How do I need
to adjust it?
[root@node01 smd]# ceph osd pool autoscale-statusPOOL