On Mon, Mar 3, 2008 at 12:08 PM, Lino Moragon <[EMAIL PROTECTED]> wrote: > Serge Dubrouski wrote: > > Configuration looks right to me, I even tested it and it worked fine > > on my test cluster. So hints are obvious: > > > > 1. Check that you really put that script on a second node and made it > > executable. > > > That was my first error, but i noticed a error message in the logfile and > corrected it. > So i can exclude this posibility > > > > 2. Nodes should be able to ping each other. That programmed in a > > "status" function. > > > What do you mean by "programmed in a status function?" From each node I can > ping the other one. It also works fine with name resolution.
I mean that "status" function in script pings each node from hostlist. If it can't ping it it fails with exit code 1. > > Do you think, it could possibly be an issue with my current version > (2.1.2-3)? No it can't because I use the same version. > With which version did you try the configurations? > when you start the Clone the first time, then both resources should start > (on node1 as well as on node2) is that correct? Right. > Could it be another configuration error outside the clone_id section in the > cib? Could you perhaps attach your whole CIB? Attached. > Thanks for your support so far, I'm very grateful. > Lino > > > > > > > > > On Mon, Mar 3, 2008 at 9:16 AM, Lino Moragon <[EMAIL PROTECTED]> wrote: > > > >> Hi, > >> > >> I'm using now the most actual xen0 stonith plugin, that Serge attachted > >> the 2008-02-28 to this thread. > >> I thought I configured everything correct but it seems that the stonith > >> clone cannot be started on my 2nd node. > >> I must admit I configured the Clone via hb_gui but I still have some > issues. > >> As for these reasons the stonith plugin only works to reset my 2nd node > >> and not vice versa. > >> The version I am using is: heartbeat-2.1.2-3 of the Centos5.1 repository. > >> > >> Has anyone a clue where the failure could be? > >> > >> To this email I attach the cib.xml > >> Below follows the cib stonith section, crm_verfy and error msgs. > >> > >> Perhaps Serge, you could give me a hint? What did you do differently > >> than me? > >> > >> I would be very glad for any hints and clues. > >> > >> Thanks in advance, > >> Lino > >> > >> node1: mysql1 > >> node2: mysql2 > >> xen-host: simulator > >> > >> <clone id="DoFencing"> > >> <instance_attributes id="DoFencing_instance_attrs"> > >> <attributes> > >> <nvpair id="DoFencing_clone_max" name="clone_max" value="2"/> > >> <nvpair id="DoFencing_clone_node_max" name="clone_node_max" > >> value="1"/> > >> <nvpair id="DoFencing_target_role" name="target_role" > >> value="started"/> > >> </attributes> > >> </instance_attributes> > >> <primitive class="stonith" type="external/xen0" > >> provider="heartbeat" id="child_DoFencing"> > >> <instance_attributes id="child_DoFencing_instance_attrs"> > >> <attributes> > >> <nvpair name="target_role" > >> id="child_DoFencing_target_role" value="started"/> > >> <nvpair name="hostlist" > >> id="5525c381-5956-4564-af3d-2bc7b547812a" value="mysql1:mysql1.cfg > >> mysql2:mysql2.cfg"/> > >> <nvpair id="65feeaf5-501f-4648-a155-83b79b587fbf" > >> name="dom0" value="simulator"/> > >> </attributes> > >> </instance_attributes> > >> </primitive> > >> </clone> > >> > >> If I use crm_verify I get the following results: > >> ============================================================= > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Processing > >> failed op (child_DoFencing:0_start_0) on mysql2 > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Handling > >> failed start for child_DoFencing:0 on mysql2 > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Processing > >> failed op (child_DoFencing:1_start_0) on mysql2 > >> crm_verify[6169]: 2008/03/03_17:06:55 WARN: unpack_rsc_op: Handling > >> failed start for child_DoFencing:1 on mysql2 > >> ============================================================= > >> > >> Furthermore I get the following Errors in my log: > >> ============================================================= > >> Mar 3 16:29:42 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM > >> operation child_DoFencing:0_start_0 (call=22, rc=1) Error unknown error > >> Mar 3 16:29:46 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM > >> operation child_DoFencing:1_start_0 (call=24, rc=1) Error unknown error > >> Mar 3 16:35:33 mysql2 crmd: [1478]: ERROR: process_lrm_event: LRM > >> operation child_DoFencing:1_start_0 (call=28, rc=1) Error unknown error > >> Mar 3 16:46:40 mysql2 crmd: [1477]: ERROR: process_lrm_event: LRM > >> operation child_DoFencing:0_start_0 (call=10, rc=1) Error unknown error > >> Mar 3 16:46:45 mysql2 crmd: [1477]: ERROR: process_lrm_event: LRM > >> operation child_DoFencing:1_start_0 (call=12, rc=1) Error unknown error > >> ============================================================== > >> > >> > >> > >> > >> > >> Serge Dubrouski schrieb: > >> > Attached. > >> > > >> > On Thu, Feb 28, 2008 at 3:35 AM, Dejan Muhamedagic <[EMAIL PROTECTED]> > wrote: > >> >> Hi Serge, > >> >> > >> >> > >> >> On Tue, Feb 26, 2008 at 09:46:14AM -0700, Serge Dubrouski wrote: > >> >> > Dejan - > >> >> > > >> >> > I found a compromise :-) Attached is a version of that plugin that > >> >> > supports following parameters: > >> >> > > >> >> > 1. hostlist. A string that has a list of controlled nodes > separated by > >> >> > space or commas. A required parameter. In a simple form its just a > >> >> > list of nodes. If one needs to use non-standard Xen configuration > he > >> >> > can use extended form of this parameter: "node1_name:config1_file > >> >> > node2_name:config2_file". If config file isn't given it defaults to > >> >> > /etc/xen/node_name.cfg > >> >> > > >> >> > 2. Dom0. Name of Dom0 Xen node. A required parameter. > >> >> > > >> >> > 3. ssh_command. SSH command that is used to ssh from DomU to Dom0. > >> >> > Defaults to "/usr/bin/ssh -q -x -n -l root". If one wants to use > SSH > >> >> > keys for higher security he needs to use this parameter. > >> >> > >> >> This is not necessary. One can setup everything needed in > >> >> ~/.ssh/config on a per host basis, i.e. key or user to connect > >> >> with. Since the plugin always runs as root, you might leave out > >> >> '-l root' as well. > >> >> > >> >> > >> >> > So in this form this plugin can be configured as a clone or as a > set > >> >> > of resources and location constraints. > >> >> > > >> >> > > >> >> > I'd be very pleased if this plugin gets its way into Linux-HA > distribution. > >> >> > > >> >> > >> >> Sure. Could you please just drop the ssh_command parameter. > >> >> > >> >> Many thanks for the contribution. > >> >> > >> >> Cheers, > >> >> > >> >> Dejan > >> >> > >> >> > >> >> > >> >> > > >> >> > On Tue, Feb 26, 2008 at 8:45 AM, Serge Dubrouski <[EMAIL > PROTECTED]> wrote: > >> >> > > > >> >> > > On Mon, Feb 25, 2008 at 4:02 PM, Dejan Muhamedagic <[EMAIL > PROTECTED]> wrote: > >> >> > > > Hi, > >> >> > > > > >> >> > > > > >> >> > > > On Mon, Feb 25, 2008 at 12:17:40PM -0700, Serge Dubrouski > wrote: > >> >> > > > > On Mon, Feb 25, 2008 at 12:10 PM, Dejan Muhamedagic > <[EMAIL PROTECTED]> wrote: > >> >> > > > > > Hi, > >> >> > > > > > > >> >> > > > > > > >> >> > > > > > On Mon, Feb 25, 2008 at 11:27:38AM -0700, Serge > Dubrouski wrote: > >> >> > > > > > > I would love to do that and already tried it. Though > we didn't come to > >> >> > > > > > > agreement on how configuration parameters should look > like. > >> >> > > > > > > >> >> > > > > > Why? Was there a discussion on the list about it? The > >> >> > > > > > configuration is a bit unusual. Other stonith agents > take named > >> >> > > > > > parameters. Though this kind of configuration also > works, I'd > >> >> > > > > > prefer something similar to the others, e.g. > >> >> > > > > > >> >> > > > > Yes there was a discussion > >> >> > > > > > http://lists.community.tummy.com/pipermail/linux-ha-dev/2007-February/ > >> >> > > > > >> >> > > > It's a long one and peters out inconclusively. > >> >> > > > > >> >> > > > > >> >> > > > > See "new stonith external plugin". The config parameter > hostslist is > >> >> > > > > actually derived from original ssh plugin. I needed to > have a full > >> >> > > > > list of all controlled nodes and preferred to have it as > one > >> >> > > > > parameter. > >> >> > > > > > >> >> > > > > > > >> >> > > > > > hostname dom0 (or xenhost) config > >> >> > > > > > > >> >> > > > > > >> >> > > > > That would work if I needed just a dom0 host, but I also > need a list > >> >> > > > > of controlled nodes and probably configuration files. > >> >> > > > > >> >> > > > That's why you can have several instances of a stonith > resource > >> >> > > > (see e.g. external/ipmi). Each of them would run with > different > >> >> > > > parameters. What I meant was: > >> >> > > > > >> >> > > > hostname: xen vm > >> >> > > > dom0: xen dom0 > >> >> > > > config: vm configuration file > >> >> > > > > >> >> > > > >> >> > > That's possible and easy to do but I'm not sure that it'll be > better. > >> >> > > Current version allows to configure a clone. New version would > require > >> >> > > configuring a separate resource for each node and creating > location > >> >> > > constraints for each of them. Per my opinion that would be more > >> >> > > complex configuration. May be I'm missing something. > >> >> > > > >> >> > > >> >> > > >> >> > > >> >> > -- > >> >> > Serge Dubrouski. > >> >> > >> >> > >> >> > >> >>> _______________________________________________ > >> >> > Linux-HA mailing list > >> >> > Linux-HA@lists.linux-ha.org > >> >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> >> > See also: http://linux-ha.org/ReportingProblems > >> >> > >> >> -- > >> >> Dejan > >> >> > >> >> > >> >> _______________________________________________ > >> >> Linux-HA mailing list > >> >> Linux-HA@lists.linux-ha.org > >> >> http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> >> See also: http://linux-ha.org/ReportingProblems > >> >> > >> > > >> > > >> > > >> > > >> > > ------------------------------------------------------------------------ > >> > > >> > _______________________________________________ > >> > Linux-HA mailing list > >> > Linux-HA@lists.linux-ha.org > >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> > See also: http://linux-ha.org/ReportingProblems > >> > >> > >> > >> _______________________________________________ > >> Linux-HA mailing list > >> Linux-HA@lists.linux-ha.org > >> http://lists.linux-ha.org/mailman/listinfo/linux-ha > >> See also: http://linux-ha.org/ReportingProblems > >> > >> > > > > > > > > > > _______________________________________________ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > -- Serge Dubrouski.
<cib admin_epoch="0" have_quorum="1" num_peers="0" cib_feature_revision="1.3" ignore_dtd="false" generated="false" crm_feature_set="2.1" epoch="281" num_updates="1" cib-last-written="Mon Mar 3 11:59:50 2008"> <configuration> <crm_config> <cluster_property_set id="cib-bootstrap-options"> <attributes> <nvpair id="cib-bootstrap-options-default_resource_stickiness" name="default-resource-stickiness" value="600"/> <nvpair id="cib-bootstrap-options-default_resource_failure_stickiness" name="default-resource-failure-stickiness" value="-520"/> <nvpair id="symmetric-cluster" name="symmetric-cluster" value="true"/> <nvpair id="stonith-enabled" name="stonith-enabled" value="true"/> <nvpair name="last-lrm-refresh" id="cib-bootstrap-options-last-lrm-refresh" value="1203565421"/> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="0.6.0-node: c94b92d550cf57217fd0292a9aa913bcf977651c"/> </attributes> </cluster_property_set> </crm_config> <nodes> <node id="ad6f19b7-228a-48b7-bae0-f95a838bde2a" uname="fc-node2" type="normal"/> <node id="b88f98c6-50f2-463a-a6eb-51abbec645a9" uname="fc-node1" type="normal"/> </nodes> <resources> <clone id="DoFencing"> <instance_attributes id="fence_attributes"> <attributes> <nvpair id="fence_clone_max" name="clone_max" value="2"/> <nvpair id="fence_clone_node_max" name="clone_node_max" value="1"/> </attributes> </instance_attributes> <primitive id="child_DoFencing" class="stonith" type="external/xen0" provider="heartbeat"> <instance_attributes id="fence_inst_attr"> <attributes> <nvpair id="xen0_hostlist" name="hostlist" value="fc-node1 fc-node2"/> <nvpair id="xen0_dom0" name="dom0" value="home"/> </attributes> </instance_attributes> </primitive> </clone> <group id="myGroup"> <instance_attributes id="myGroup_instance_attrs"> <attributes/> </instance_attributes> <primitive class="ocf" type="IPaddr" provider="heartbeat" id="myIP"> <instance_attributes id="myIP_attributes"> <attributes> <nvpair id="myIP_ip" name="ip" value="192.168.1.130"/> </attributes> </instance_attributes> <operations> <op id="63460aec-8759-4a35-a41c-0e402d5409a0" name="monitor" interval="30s" timeout="30s"/> <op id="75bb6f39-b41c-4837-8714-9fc2305fa4c0" name="start" interval="0s" timeout="30s"/> <op id="9c694b65-9c32-43d5-8df2-615dd9dbe56e" name="stop" interval="0s" timeout="30s"/> </operations> <instance_attributes id="myIP"> <attributes/> </instance_attributes> </primitive> <primitive class="ocf" type="pgsql" provider="heartbeat" id="myPgsql"> <instance_attributes id="myPgsql_instance_attrs"> <attributes> <nvpair id="pgsql_ctl_opt" name="ctl_opt" value="-w"/> </attributes> </instance_attributes> <operations> <op id="pgsql_monitor" name="monitor" interval="30s" timeout="30s"/> <op id="pgsql_start" name="start" interval="0s" timeout="30s"/> <op id="pgsal_stop" name="stop" interval="0s" timeout="30s"/> </operations> <instance_attributes id="myPgsql"> <attributes/> </instance_attributes> </primitive> <instance_attributes id="myGroup"> <attributes> <nvpair id="myGroup-is_managed" name="is_managed" value="true"/> </attributes> </instance_attributes> </group> </resources> <constraints> <rsc_location id="primNode" rsc="myGroup"> <rule id="prefered_primNode" score="1000"> <expression attribute="#uname" id="906247e1-1d96-4a63-a80b-13d103d1b31c" operation="eq" value="fc-node1"/> </rule> </rsc_location> <rsc_location id="PGSQL:connected" rsc="myGroup"> <rule id="PGSQL:connected:rule" score="-INFINITY" boolean_op="or"> <expression id="PGSQL:connected:expr:undefined" attribute="pingd" operation="not_defined"/> <expression id="PGSQL:connected:expr:zero" attribute="pingd" operation="lte" value="0"/> </rule> </rsc_location> </rsc_location> </constraints> </configuration> </cib>
_______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems