Like most disk redundancy systems, the concern usually is the amount of
time it takes to recover, wherein you are vulnerable to another failure. I
would assume that is also the concern here.


On Wed, Jul 1, 2015 at 5:54 PM, Nate Curry <cu...@mosaicatm.com> wrote:

> 4TB is too much to lose?  Why would it matter if you lost one 4TB with the
> redundancy?  Won't it auto recover from the disk failure?
>
> Nate Curry
> On Jul 1, 2015 6:12 PM, "German Anders" <gand...@despegar.com> wrote:
>
>> I would probably go with less size osd disks, 4TB is to much to loss in
>> case of a broken disk, so maybe more osd daemons with less size, maybe 1TB
>> or 2TB size. 4:1 relationship is good enough, also i think that 200G disk
>> for the journals would be ok, so you can save some money there, the osd's
>> of course configured them as a JBOD, don't use any RAID under it, and use
>> two different networks for public and cluster net.
>>
>> *German*
>>
>> 2015-07-01 18:49 GMT-03:00 Nate Curry <cu...@mosaicatm.com>:
>>
>>> I would like to get some clarification on the size of the journal disks
>>> that I should get for my new Ceph cluster I am planning.  I read about the
>>> journal settings on
>>> http://ceph.com/docs/master/rados/configuration/osd-config-ref/#journal-settings
>>> but that didn't really clarify it for me that or I just didn't get it.  I
>>> found in the Learning Ceph Packt book it states that you should have one
>>> disk for journalling for every 4 OSDs.  Using that as a reference I was
>>> planning on getting multiple systems with 8 x 6TB inline SAS drives for
>>> OSDs with two SSDs for journalling per host as well as 2 hot spares for the
>>> 6TB drives and 2 drives for the OS.  I was thinking of 400GB SSD drives but
>>> am wondering if that is too much.  Any informed opinions would be
>>> appreciated.
>>>
>>> Thanks,
>>>
>>> *Nate Curry*
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to