> Op 13 aug. 2016 om 08:58 heeft Georgios Dimitrakakis <gior...@acmac.uoc.gr> > het volgende geschreven: > > >>> Op 13 aug. 2016 om 03:19 heeft Bill Sharer het volgende geschreven: >>> >>> If all the system disk does is handle the o/s (ie osd journals are >>> on dedicated or osd drives as well), no problem. Just rebuild the >>> system and copy the ceph.conf back in when you re-install ceph. >>> Keep a spare copy of your original fstab to keep your osd filesystem >>> mounts straight. >> >> With systems deployed with ceph-disk/ceph-deploy you no longer need a >> fstab. Udev handles it. >> >>> Just keep in mind that you are down 11 osds while that system drive >>> gets rebuilt though. It's safer to do 10 osds and then have a >>> mirror set for the system disk. >> >> In the years that I run Ceph I rarely see OS disks fail. Why bother? >> Ceph is designed for failure. >> >> I would not sacrifice a OSD slot for a OS disk. Also, let's say a >> additional OS disk is €100. >> >> If you put that disk in 20 machines that's €2.000. For that money >> you can even buy a additional chassis. >> >> No, I would run on a single OS disk. It fails? Let it fail. Re-install >> and you're good again. >> >> Ceph makes sure the data is safe. >> > > Wido, > > can you elaborate a little bit more on this? How does CEPH achieve that? Is > it by redundant MONs? >
No, Ceph replicates over hosts by default. So you can loose a host and the other ones will have copies. > To my understanding the OSD mapping is needed to have the cluster back. In > our setup (I assume in others as well) that is stored in the OS > disk.Furthermore, our MONs are running on the same host as OSDs. So if the OS > disk fails not only we loose the OSD host but we also loose the MON node. Is > there another way to be protected by such a failure besides additional MONs? > Aha, MON on the OSD host. I never recommend that. Try to use dedicated machines with a good SSD for MONs. Technically you can run the MON on the OSD nodes, but I always try to avoid it. It just isn't practical when stuff really goes wrong. Wido > We recently had a problem where a user accidentally deleted a volume. Of > course this has nothing to do with OS disk failure itself but we 've been in > the loop to start looking for other possible failures on our system that > could jeopardize data and this thread got my attention. > > > Warmest regards, > > George > > >> Wido >> >> Bill Sharer >> >>> On 08/12/2016 03:33 PM, Ronny Aasen wrote: >>> >>>> On 12.08.2016 13:41, Félix Barbeira wrote: >>>> >>>> Hi, >>>> >>>> I'm planning to make a ceph cluster but I have a serious doubt. At >>>> this moment we have ~10 servers DELL R730xd with 12x4TB SATA >>>> disks. The official ceph docs says: >>>> >>>> "We recommend using a dedicated drive for the operating system and >>>> software, and one drive for each Ceph OSD Daemon you run on the >>>> host." >>>> >>>> I could use for example 1 disk for the OS and 11 for OSD data. In >>>> the operating system I would run 11 daemons to control the OSDs. >>>> But...what happen to the cluster if the disk with the OS fails?? >>>> maybe the cluster thinks that 11 OSD failed and try to replicate >>>> all that data over the cluster...that sounds no good. >>>> >>>> Should I use 2 disks for the OS making a RAID1? in this case I'm >>>> "wasting" 8TB only for ~10GB that the OS needs. >>>> >>>> In all the docs that i've been reading says ceph has no unique >>>> single point of failure, so I think that this scenario must have a >>>> optimal solution, maybe somebody could help me. >>>> >>>> Thanks in advance. >>>> >>>> -- >>>> >>>> Félix Barbeira. >>> if you do not have dedicated slots on the back for OS disks, then i >>> would recomend using SATADOM flash modules directly into a SATA port >>> internal in the machine. Saves you 2 slots for osd's and they are >>> quite reliable. you could even use 2 sd cards if your machine have >>> the internal SD slot >>> >>> >> http://www.dell.com/downloads/global/products/pedge/en/poweredge-idsdm-whitepaper-en.pdf >>> [1] >>> >>> kind regards >>> Ronny Aasen >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com [2] >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com [3] >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-u >> ph.com >> http://li >> >>> i/ceph-users-ceph.com >> >> >> Links: >> ------ >> [1] >> http://www.dell.com/downloads/global/products/pedge/en/poweredge-idsdm-whitepaper-en.pdf >> [2] mailto:ceph-users@lists.ceph.com >> [3] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> [4] mailto:bsha...@sharerland.com > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com