Our cluster here having troubles is primarily for object storage, and
somewhere around 650M objects and 600T. Majority of objects being small
jpgs, large objects are big movie .ts files and .mp4.

This was upgraded from jewel on xenial last month.majority of bugs are
ceph-osd on SSDs for us. We've got bugs on filestore and bluestore at this
point with ssd journals, colocated wal/db, nvme wal/db, you name it. We're
removing the SSDs from the cluster right now. They just keep failing and
causing downtime. Rather a slow cluster than an unusable one. A crashy osd
is a gone osd nowadays.

E

On Thu, Nov 16, 2017 at 7:22 PM Konstantin Shalygin <k0...@k0ste.ru> wrote:

> > My cluster (55 OSDs) runs 12.2.x since the release, and bluestore too
> > All good so far
> This is cleanly deployed cluster or upgrade from some version?
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to