Hi all,

now that host masks seem to work, could somebody please shed some light at the 
relative priority of these settings:

ceph config set osd memory_target X
ceph config set osd/host:A memory_target Y
ceph config set osd/class:B memory_target Z

Which one wins for an OSD on host A in class B?

Similar for an explicit ID. The expectation is that a setting for OSD.ID always 
wins. Then the masked values, then the generic osd setting, then the globals 
and last the defaults. The relative precedence of masked values is not defined 
anywhere, nor the precedence in general.

This is missing in even the latest docs: 
https://docs.ceph.com/en/quincy/rados/configuration/ceph-conf/#sections-and-masks
 . Would be great if someone could add this.

Thanks!
=================
Frank Schilder
AIT Risø Campus
Bygning 109, rum S14

________________________________________
From: Luis Domingues <luis.doming...@proton.ch>
Sent: Monday, July 17, 2023 9:36 AM
To: Sridhar Seshasayee
Cc: Mark Nelson; ceph-users@ceph.io
Subject: [ceph-users] Re: OSD memory usage after cephadm adoption

It looks indeed to be that bug that I hit.

Thanks.

Luis Domingues
Proton AG


------- Original Message -------
On Monday, July 17th, 2023 at 07:45, Sridhar Seshasayee <ssesh...@redhat.com> 
wrote:


> Hello Luis,
>
> Please see my response below:
>
> But when I took a look on the memory usage of my OSDs, I was below of that
>
> > value, by quite a bite. Looking at the OSDs themselves, I have:
> >
> > "bluestore-pricache": {
> > "target_bytes": 4294967296,
> > "mapped_bytes": 1343455232,
> > "unmapped_bytes": 16973824,
> > "heap_bytes": 1360429056,
> > "cache_bytes": 2845415832
> > },
> >
> > And if I get the running config:
> > "osd_memory_target": "4294967296",
> > "osd_memory_target_autotune": "true",
> > "osd_memory_target_cgroup_limit_ratio": "0.800000",
> >
> > Which is not the value I observe from the config. I have 4294967296
> > instead of something around 7219293672. Did I miss something?
>
> This is very likely due to https://tracker.ceph.com/issues/48750. The fix
> was recently merged into
> the main branch and should be backported soon all the way to pacific.
>
> Until then, the workaround would be to set the osd_memory_target on each
> OSD individually to
> the desired value.
>
> -Sridhar
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to