Easiest way I know would be to use
$ ceph tell osd.X compact

This is what cures that whenever I have metadata spillover.

Reed

> On Mar 2, 2020, at 3:32 AM, Stefan Priebe - Profihost AG 
> <s.pri...@profihost.ag> wrote:
> 
> Hello,
> 
> i added a db device to my osds running nautilus. The DB data migratet
> over some days from the hdd to ssd (db device).
> 
> But now it seems all are stuck at:
> # ceph health detail
> HEALTH_WARN BlueFS spillover detected on 8 OSD(s)
> BLUEFS_SPILLOVER BlueFS spillover detected on 8 OSD(s)
>     osd.0 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.1 spilled over 3.4 MiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.2 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.3 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.4 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.5 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.6 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
>     osd.7 spilled over 128 KiB metadata from 'db' device (12 GiB used
> of 185 GiB) to slow device
> 
> any idea how to force ceph to move the last 128kb to the db device?
> 
> Greets,
> Stefan
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to