Easiest way I know would be to use $ ceph tell osd.X compact This is what cures that whenever I have metadata spillover.
Reed > On Mar 2, 2020, at 3:32 AM, Stefan Priebe - Profihost AG > <s.pri...@profihost.ag> wrote: > > Hello, > > i added a db device to my osds running nautilus. The DB data migratet > over some days from the hdd to ssd (db device). > > But now it seems all are stuck at: > # ceph health detail > HEALTH_WARN BlueFS spillover detected on 8 OSD(s) > BLUEFS_SPILLOVER BlueFS spillover detected on 8 OSD(s) > osd.0 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.1 spilled over 3.4 MiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.2 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.3 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.4 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.5 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.6 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > osd.7 spilled over 128 KiB metadata from 'db' device (12 GiB used > of 185 GiB) to slow device > > any idea how to force ceph to move the last 128kb to the db device? > > Greets, > Stefan > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io