I think a few other things that could help would be `ceph osd df tree` which 
will show the hierarchy across different crush domains.
And if you’re doing something like erasure coded pools, or something other than 
replication 3, maybe `ceph osd crush rule dump` may provide some further 
context with the tree output.

Also, the cluster is running Luminous (12) which went EOL 3 years ago tomorrow 
<https://docs.ceph.com/en/latest/releases/index.html#archived-releases>.
So there are also likely a good bit of improvements all around under the hood 
to be gained by moving forward from Luminous.
Though, I would say take care of the scrub errors prior to doing any major 
upgrades, as well as checking your upgrade path (can only upgrade two releases 
at a time, if you have filestore OSDs, etc).

-Reed

> On Feb 28, 2023, at 11:12 AM, Dave Ingram <d...@adaptable.sh> wrote:
> 
> There is a
> lot of variability in drive sizes - two different sets of admins added
> disks sized between 6TB and 16TB and I suspect this and imbalanced
> weighting is to blame.
> 
> CEPH OSD DF:
> 
> (not going to paste that all in here): https://pastebin.com/CNW5RKWx
> 
> What else am I missing in terms of what to share with you all?
> 
> Thanks all,
> -Dave
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io


Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to