On Fri, Apr 20, 2007 at 10:25:04AM -0400, Bjorn Oglefjorn wrote: > If it seems counter intuitive, think of it like this: > * test-1_DRAC is the DRAC installed in the chassis of > test-1.domainwhich has an address of > test-1.drac.domain
No, actually it's not counter intuitive. > Then look here: > <rsc_location id="test-1_DRAC_location" rsc="test-1_DRAC"> > <rule id="no_self_run_test-1_DRAC" score="-INFINITY"> > <expression attribute="#uname" > id="no_self_run_test-1_DRAC_expr_1" operation="eq" value="test-1.domain"/> > </rule> > </rsc_location> > <rsc_location id="test-2_DRAC_location" rsc="test-2_DRAC"> > <rule id="no_self_run_test-2_DRAC" score="-INFINITY"> > <expression attribute="#uname" > id="no_self_run_test-2_DRAC_expr_1" operation="eq" value="test-2.domain"/> > </rule> > </rsc_location> > > In other words, test-1_DRAC should never run on test-1.domain and > test-2_DRAC should never run on test-2.domain. > > Once again: > > [EMAIL PROTECTED] ~]# stonith -t external/drac4 > DRAC_ADDR=test-2.drac.domainDRAC_LOGIN=root DRAC_PASSWD=******** -lS > stonith: external/drac4 device OK. > test-2.drac.domain This is wrong. It should output the name of the host which it can manage, i.e. test-2.domain. > [EMAIL PROTECTED] ~]# stonith -t external/drac4 > DRAC_ADDR=test-1.drac.domainDRAC_LOGIN=root DRAC_PASSWD=******** -lS > stonith: external/drac4 device OK. > test-1.drac.domain Same here. > > --BO > > On 4/19/07, Dejan Muhamedagic <[EMAIL PROTECTED]> wrote: > > > >On Tue, Apr 17, 2007 at 03:55:07PM -0400, Bjorn Oglefjorn wrote: > >> Alan, what is the list operation? The node names are always FQDNs and > >> always match. > > > >Do they? > > > >>From your CIB: > > > > <primitive id="test-1_DRAC" class="stonith" type="external/drac4" > >provider="heartbeat"> > > <operations> > > <op id="test-1_DRAC_reset" name="reset" timeout="3min" > >prereq="nothing"/> > > </operations> > > <instance_attributes id="test-1_DRAC_inst_attr"> > > <attributes> > > <nvpair id="test-1_DRAC_attr_0" name="DRAC_ADDR" value=" > >test-1.drac.domain"/> > > <nvpair id="test-1_DRAC_attr_1" name="DRAC_LOGIN" > >value="root"/> > > <nvpair id="test-1_DRAC_attr_2" name="DRAC_PASSWD" > >value="********"/> > > </attributes> > > </instance_attributes> > > </primitive> > > > >Shouldn't attr_0 specify test-2.drac instead of test-1? > > > >BTW, pulling network cables from a host results in a split brain. > >That's to be avoided at all costs, that's why you have redundant > >connections within the cluster. When that happens, both nodes will > >try to kill each other. A better way to test for a dead node is to > > > ># killall -9 heartbeat > > > >(that's for linux; use kill -9 <pid> to kill the master process on > >other platforms) > > > >> --BO > >> > >> On 4/17/07, Alan Robertson <[EMAIL PROTECTED]> wrote: > >> > > >> >Andrew Beekhof wrote: > >> >> On 4/17/07, Bjorn Oglefjorn <[EMAIL PROTECTED]> wrote: > >> >>> I know that my plugin is getting called because of the logging that > >the > >> >>> plugin does. > >> >> > >> >> do we get to see that logging at all? preferably in the context of > >> >> the other log messages > >> >> > >> >>> That said, I also know my plugin is not receiving any 'reset' > >> >>> operation request from heartbeat. If you see below, request actions > >> >are > >> >>> logged. The only actions logged when node failure is simulated are: > >> >>> getconfignames, status, and gethosts, in that order. We should also > >> >see > >> >>> getinfo-devid and reset operations logged, but they are never > >present. > >> > > >> >I would assume that you're getting called with the list operation, but > >> >not afterwards. > >> > > >> >If that's the case, then that means that for some reason not obvious to > >> >me the stonith daemon doesn't think the names you gave it match > >> >the host names it is being asked to reset. > >> > > >> >-- > >> > Alan Robertson <[EMAIL PROTECTED]> > >> > > >> >"Openness is the foundation and preservative of friendship... Let me > >> >claim from you at all times your undisguised opinions." - William > >> >Wilberforce > >> >_______________________________________________ > >> >Linux-HA mailing list > >> >Linux-HA@lists.linux-ha.org > >> >http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> >See also: http://linux-ha.org/ReportingProblems > >> > > >> _______________________________________________ > >> Linux-HA mailing list > >> Linux-HA@lists.linux-ha.org > >> http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> See also: http://linux-ha.org/ReportingProblems > > > >-- > >Dejan > >_______________________________________________ > >Linux-HA mailing list > >Linux-HA@lists.linux-ha.org > >http://lists.linux-ha.org/mailman/listinfo/linux-ha > >See also: http://linux-ha.org/ReportingProblems > > > _______________________________________________ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems -- Dejan _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems