Have had some good experiences with ST10000NM0156-2AA111

Also running with EC but using RBD for slow storage for VM's, have had good
performance for what id expect for 10TB drive's and EC.

I would definitely say getting helium drives helps VS standard air filled
when you get to the 8TB+ drives.

On Wed, Jan 16, 2019 at 5:34 PM Götz Reinicke <
goetz.reini...@filmakademie.de> wrote:

> Dear Ceph users,
>
> I’d like to get some feedback for the following thought:
>
> Currently I run some 24*4TB bluestore OSD nodes. The main focus is on
> storage space over IOPS.
>
> We use erasure code and cephfs, and things look good right now.
>
> The „but“ is, I do need more disk space and don’t have so much more rack
> space available, so I was thinking of adding some 8TB or even 12TB OSDs
> and/or exchange over time 4TB OSDs with bigger disks.
>
> My question is: How are your experiences with the current >=8TB SATA disks
> are some very bad models out there which I should avoid?
>
> The current OSD nodes are connected by 4*10Gb bonds, so for
> replication/recovery speed is a 24 Chassis with bigger disks useful, or
> should I go with smaller chassis? Or dose the chassi sice does not matter
> at all that much in my setup.
>
> I know, EC is quit computing intense, so may be bigger disks hav also
> there an impact?
>
> Lot’s of questions, may be you can help answering some.
>
>         Best regards and Thanks a lot for feedback . Götz
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to