Thanks for reporting, Nick - I've seen the same thing and thought I was just crazy.
Chris On Wed, Feb 11, 2015 at 6:48 AM, Nick Fisk <n...@fisk.me.uk> wrote: > Hi David, > > > > I have had a few weird issues when shutting down a node, although I can > replicate it by doing a “stop ceph-all” as well. It seems that OSD failure > detection takes a lot longer when a monitor goes down at the same time, > sometimes I have seen the whole cluster grind to a halt for several minutes > before it works out whats happened. > > > > If I stop the either role and wait for it to be detected as failed and > then do the next role, I don’t see the problem. So it might be something to > keep in mind when doing maintenance. > > > > Nick > > > > *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf > Of *David Graham > *Sent:* 10 February 2015 17:07 > *To:* ceph-us...@ceph.com > *Subject:* [ceph-users] combined ceph roles > > > > Hello, I'm giving thought to a minimal footprint scenario with full > redundancy. I realize it isn't ideal--and may impact overall performance > -- but wondering if the below example would work, supported, or known to > cause issue? > > Example, 3x hosts each running: > -- OSD's > -- Mon > -- Client > > I thought I read a post a while back about Client+OSD on the same host > possibly being an issue -- but i am having difficulty finding that > reference. > > I would appreciate if anyone has insight into such a setup, > > thanks! > > > > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com