If you have a small cluster, without host redundancy, you are still
able to configure this in Ceph to be handled correctly by adding a
drive failure domain between host and OSD level. So yes you need to
change more then just failure-domain=OSD, as this would be a problem.
However it is absolutely the same as to having multiple OSDs per NVMe
as some people do it.

--
Martin Verges
Managing director

Mobile: +49 174 9335695
E-Mail: martin.ver...@croit.io
Chat: https://t.me/MartinVerges

croit GmbH, Freseniusstr. 31h, 81247 Munich
CEO: Martin Verges - VAT-ID: DE310638492
Com. register: Amtsgericht Munich HRB 231263

Web: https://croit.io
YouTube: https://goo.gl/PGE1Bx

Am Sa., 13. März 2021 um 13:11 Uhr schrieb Marc <m...@f1-outsourcing.eu>:
>
> > Well, if you run with failure-domain=host, then if it says "I have 8
> > 14TB drives and one failed" or "I have 16 7TB drives and two failed"
> > isn't going to matter much in terms of recovery, is it?
> > It would mostly matter for failure-domain=OSD, otherwise it seems about
> > equal.
>
> Yes, but especially in small clusters, people are changing the failure domain 
> to osd to be able to use EC (like I have ;))
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to