Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Lars Fenneberg
Hi Konstantin!

Quoting Konstantin Shalygin (k0...@k0ste.ru):

> >Is there any recommandation of how many osds a single flash device can
> >serve? The optane ones can do 2000MB/s write + 500.000 iop/s.
> 
> Any sizes of db, except 3/30/300 is useless.

I have this from Mattia Belluco in my notes which suggests that twice the
amount is best:

> Following some discussions we had at the past Cephalocon I beg to differ
> on this point: when RocksDB needs to compact a layer it rewrites it
> *before* deleting the old data; if you'd like to be sure you db does not
> spill over to the spindle you should allocate twice the size of the
> biggest layer to allow for compaction. I guess ~60 GB would be the sweet
> spot assuming you don't plan to mess with size and multiplier of the
> rocksDB layers and don't want to go all the way to 600 GB (300 GB x2)

Source is 
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2019-May/035086.html.

And apart from the RocksDB pecularities the actual use case also needs to be
considered.  Lots of small files on a CephFS will require more DB space than
mainly big files as Paul states in the same thread.

Cheers,
LF.
-- 
Lars Fenneberg, l...@elemental.net
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Konstantin Shalygin

i'm plannung to split the block db to a seperate flash device which i
also would like to use as an OSD for erasure coding metadata for rbd
devices.

If i want to use 14x 14TB HDDs per Node
https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing

recommends a minimum size of 140GB per 14TB HDD.

Is there any recommandation of how many osds a single flash device can
serve? The optane ones can do 2000MB/s write + 500.000 iop/s.


Any sizes of db, except 3/30/300 is useless.

How much OSD's per NVMe - quantity of OSD's that you can lose once at time.



k

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Xiaoxi Chen
One tricky thing is each layer of RocksDB is 100% on SSD or 100% on HDD,
so either you need to tweak the rocksdb configuration , or there will be a
huge waste,  e.g  20GB DB partition makes no difference compared to a 3GB
one (under default rocksdb configuration)

Janne Johansson  于2020年1月14日周二 下午4:43写道:

> (sorry for empty mail just before)
>
>
>> i'm plannung to split the block db to a seperate flash device which i
>>> also would like to use as an OSD for erasure coding metadata for rbd
>>> devices.
>>>
>>> If i want to use 14x 14TB HDDs per Node
>>>
>>> https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing
>>>
>>> recommends a minimum size of 140GB per 14TB HDD.
>>>
>>> Is there any recommandation of how many osds a single flash device can
>>> serve? The optane ones can do 2000MB/s write + 500.000 iop/s.
>>>
>>
>>
> I think many ceph admins are more concerned with having many drives
> co-using the same DB drive, since if the DB drive fails, it also means all
> OSDs are lost at the same time.
> Optanes and decent NVMEs are probably capable of handling tons of HDDs, so
> that the bottleneck ends up being somewhere else, but the failure scenarios
> are a bit scary if the whole host is lost just by that one DB device acting
> up.
>
> --
> May the most significant bit of your life be positive.
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Janne Johansson
(sorry for empty mail just before)


> i'm plannung to split the block db to a seperate flash device which i
>> also would like to use as an OSD for erasure coding metadata for rbd
>> devices.
>>
>> If i want to use 14x 14TB HDDs per Node
>>
>> https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing
>>
>> recommends a minimum size of 140GB per 14TB HDD.
>>
>> Is there any recommandation of how many osds a single flash device can
>> serve? The optane ones can do 2000MB/s write + 500.000 iop/s.
>>
>
>
I think many ceph admins are more concerned with having many drives
co-using the same DB drive, since if the DB drive fails, it also means all
OSDs are lost at the same time.
Optanes and decent NVMEs are probably capable of handling tons of HDDs, so
that the bottleneck ends up being somewhere else, but the failure scenarios
are a bit scary if the whole host is lost just by that one DB device acting
up.

-- 
May the most significant bit of your life be positive.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Janne Johansson
Den mån 13 jan. 2020 kl 08:09 skrev Stefan Priebe - Profihost AG <
s.pri...@profihost.ag>:

> Hello,
>
> i'm plannung to split the block db to a seperate flash device which i
> also would like to use as an OSD for erasure coding metadata for rbd
> devices.
>
> If i want to use 14x 14TB HDDs per Node
>
> https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing
>
> recommends a minimum size of 140GB per 14TB HDD.
>
> Is there any recommandation of how many osds a single flash device can
> serve? The optane ones can do 2000MB/s write + 500.000 iop/s.
>
> Greets,
> Stefan
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


-- 
May the most significant bit of your life be positive.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] block db sizing and calculation

2020-01-14 Thread Stefan Priebe - Profihost AG
Hello,

does anybody have real live experience with externel block db?

Greets,
Stefan
Am 13.01.20 um 08:09 schrieb Stefan Priebe - Profihost AG:
> Hello,
> 
> i'm plannung to split the block db to a seperate flash device which i
> also would like to use as an OSD for erasure coding metadata for rbd
> devices.
> 
> If i want to use 14x 14TB HDDs per Node
> https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing
> 
> recommends a minimum size of 140GB per 14TB HDD.
> 
> Is there any recommandation of how many osds a single flash device can
> serve? The optane ones can do 2000MB/s write + 500.000 iop/s.
> 
> Greets,
> Stefan
> 
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] block db sizing and calculation

2020-01-12 Thread Stefan Priebe - Profihost AG
Hello,

i'm plannung to split the block db to a seperate flash device which i
also would like to use as an OSD for erasure coding metadata for rbd
devices.

If i want to use 14x 14TB HDDs per Node
https://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/#sizing

recommends a minimum size of 140GB per 14TB HDD.

Is there any recommandation of how many osds a single flash device can
serve? The optane ones can do 2000MB/s write + 500.000 iop/s.

Greets,
Stefan
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com