To set these setting during runtime I use the following commands from my admin node:
ceph tell osd.* injectargs '--osd-max-backfills 1' ceph tell osd.* injectargs '--osd-recovery-max-active 1' ceph tell osd.* injectargs '--osd-op-queue-cut-off high' Right? On Wed, Aug 19, 2020 at 11:08 AM Stefan Kooman <ste...@bit.nl> wrote: > > On 2020-08-19 04:05, norman wrote: > > Hans, > > > > I made a big change in my staging cluster before, I set a pool pg_num > > from 8 to 2048, it cased the cluster available for a long time :( > > I doubt it will have that big an effect this time. The change from 8 -> > 2048 is way bigger than "just" a doubling to 4096. It of course depends > on how much data has been added to the pool in the meantime. But if the > PGs hold less data per PG, the impact is probably lower. > > It's always a tradeoff between recovery speed and decent client IO > performance, and highly dependent per use case, cluster size, hardware > specifications like disk type, amount of RAM / CPU and networking. > > For some clusters the values you have used might by no big deal (if you > have hundreds of nodes) but for small(er) clusters this can have a big > impact. Might, if you have NVMe and loads of memory and CPU you might > even get away with it. > > So it's best to start with low(est) possible recovery/backfill settings > and slowly scaling up. Having metrics of your cluster and client VMs > (perceived latency on VMs for example) will be crucial to put a number > on this. Having a baseline of the performance of the cluster will help > to decide when things get out of hand. It will also help to identify the > time the cluster is less busy, and use that time window to perform > maintenance (i.e. not when backups are active, which tend to stress > clusters a lot). > > Gr. Stefan > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io