Thank you so much! That solved the problem.
Timothy Carr wrote: > Hi mark, > > There is a hostcache file which yuo can remove located under /var/. > > Stop Heartbeat, make a backup of your hostcache file then remove the > hostcache file. STart heartbeat and have a look again. > > Having renamed your machine names will cause problems with heartbeat. > > Tim > > > On Fri, Oct 2, 2009 at 9:34 AM, Mark Hunting <m...@netexpo.nl> wrote: > > >> Mark Hunting wrote: >> >>> Dejan Muhamedagic wrote: >>> >>> >>>> Hi, >>>> >>>> On Thu, Oct 01, 2009 at 04:45:45PM +0200, Mark Hunting wrote: >>>> >>>> >>>> >>>>> Sorry for not mentioning I use Heartbeat 2.1.3 from the Debian Lenny >>>>> repository, in crm configuration. >>>>> >>>>> Mark Hunting wrote: >>>>> >>>>> >>>>> >>>>>> Hi, >>>>>> >>>>>> I have set up a 3-node cluster. Works perfectly, but when I shut one >>>>>> node down the other two lose quorum, and shut down their resources (!) >>>>>> because no-quorum-policy is set to 'stop' like it should. >>>>>> I have no idea why the quorum is lost, this really should not happen >>>>>> >> as >> >>>>>> the remaining two nodes are still the majority. crm_mon shows them >>>>>> online and they can talk to each other. Only the quorum is lost, >>>>>> have_quorum is "false" until the third node comes up again. >>>>>> Can anybody tell me how this is possible, or give me some command that >>>>>> can help me investigate this? >>>>>> >>>>>> >>>>>> >>>> ccm_tool (or similar, can't recall the name exactly) can show you >>>> what a node thinks its partition looks like. Otherwise, look at >>>> the ccm lines in the logs, though they may be really hard to >>>> figure out. >>>> >>>> Thanks, >>>> >>>> Dejan >>>> >>>> >>> Thanks a lot! It just came to my mind that I changed the three node >>> names today in ha.cf, and this problem started to occur afterwards. I >>> think the cluster still remembers the three old names next to the new >>> ones. I guess it now 'thinks' it has six nodes instead of three, and >>> that may be an explanation for this behaviour I'm seeing (although then >>> with 3 of the 6 nodes online it also shouldn't get a quorum imo, but it >>> does). crm_admin shows only 3 nodes however, that's a bit strange. I >>> can't access the cluster right now, but I'll try to figure out more >>> tomorrow. There should be a way to force the removal of the old node >>> names (ideas anyone?) >>> >> I know a bit more now. The cluster thinks it has 4 nodes instead of 3. I >> see this in my logs: >> ccm: [5131]: debug: total_node_count=4, total_quorum_votes=400 >> But there are really only 3 nodes. Crmadmin, ccm_tool and the xml output >> from cibadmin all only show my existing 3 nodes. So I have no idea >> where this total_node_count of 4 comes from. How can I let Heartbeat >> stop thinking it has 4 nodes? >> _______________________________________________ >> Linux-HA mailing list >> Linux-HA@lists.linux-ha.org >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> See also: http://linux-ha.org/ReportingProblems >> >> > > > > _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems