Better way is increase osd set-full-ratio slightly (.97) and then remove
buckets.

-AmitG

On Wed, 30 Jan 2019, 21:30 Paul Emmerich, <paul.emmer...@croit.io> wrote:

> Quick and dirty solution: take the full OSD down to issue the deletion
> command ;)
>
> Better solutions: temporarily incrase the full limit (ceph osd
> set-full-ratio) or reduce the OSD's reweight (ceph osd reweight)
>
>
> Paul
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 München
> www.croit.io
> Tel: +49 89 1896585 90
>
> On Wed, Jan 30, 2019 at 11:56 AM Fabio - NS3 srl <fa...@ns3.it> wrote:
> >
> > Hello guys,
> > i have a Ceph with a full S3
> >
> > ~# ceph health detail
> > HEALTH_ERR 1 full osd(s); 1 near full osd(s)
> > osd.2 is full at 95%
> > osd.5 is near full at 85%
> >
> >
> > I want to delete some bucket but when i tried to show list bucket
> >
> >
> > ~# radosgw-admin bucket list
> > 2019-01-30 11:41:47.933621 7f467a9d0780  0 client.3967227.objecter
> FULL, paused modify 0x2aaf410 tid 8
> >
> > the command remains blocked ...no prompt.
> >
> > Solutions .... as well as adding an OSD?
> >
> > Many thankssss
> > --
> > Fabio
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to