Thank you Robin.

Looking at the video it doesn't seem like a fix is anywhere near ready.
Am I correct in concluding that Ceph is not the right tool for my use-case?

Cheers,
Christian

On Oct 3 2019, at 6:07 am, Robin H. Johnson <robb...@gentoo.org> wrote:
> On Wed, Oct 02, 2019 at 01:48:40PM +0200, Christian Pedersen wrote:
> > Hi Martin,
> >
> > Even before adding cold storage on HDD, I had the cluster with SSD only. 
> > That also could not keep up with deleting the files.
> > I am no where near I/O exhaustion on the SSDs or even the HDDs.
>
> Please see my presentation from Cephalic on 2019 about RGW S3 where I
> touch on slowness in Lifecycle processing and deletion.
>
> The efficiency of the code is very low: it requires a full scan of
> the bucket index every single day. Depending on the traversal order
> (unordered listing helps), this might mean it takes a very long time to
> find the items that can be deleted, and even when it gets to them, it's
> bound by the deletion time, which is also slow (that the head of the
> objects is a synchronous deletion in many cases, while the tails are
> async garbage-collected).
>
> Fixing this isn't trivial: either you have to scan the entire bucket, or
> you have to maintain a secondary index in insertion-order for EACH
> prefix in a lifecycle policy.
>
> --
> Robin Hugh Johnson
> Gentoo Linux: Dev, Infra Lead, Foundation Treasurer
> E-Mail : robb...@gentoo.org
> GnuPG FP : 11ACBA4F 4778E3F6 E4EDF38E B27B944E 34884E85
> GnuPG FP : 7D0B3CEB E9B85B1F 825BCECF EE05E6F6 A48F6136
>

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to