I defined two ping clone resources, to be used independently by different resources:
<clone id="clone-pingd-internal"> <primitive id="pingd-internal" provider="pacemaker" class="ocf" type="pingd"> <instance_attributes id="pingd-internal-ia"> <nvpair id="pingd-internal-ia01" name="name" value="pingd-internal"/> <nvpair id="pingd-internal-ia02" name="dampen" value="5s"/> <nvpair id="pingd-internal-ia03" name="multiplier" value="1000"/> <nvpair id="pingd-internal-ia04" name="host_list" value="172.17.32.23 192.168.132.23"/> </instance_attributes> </primitive> </clone> <clone id="clone-pingd-external"> <primitive id="pingd-external" provider="pacemaker" class="ocf" type="pingd"> <instance_attributes id="pingd-external-ia"> <nvpair id="pingd-external-ia01" name="name" value="pingd-external"/> <nvpair id="pingd-external-ia02" name="dampen" value="5s"/> <nvpair id="pingd-external-ia03" name="multiplier" value="1000"/> <nvpair id="pingd-external-ia04" name="host_list" value="195.244.97.241"/> </instance_attributes> </primitive> </clone> I defined a constraint for a resource so that it depends on pingd-internal <constraints> <rsc_location id="hbtest1b-connectivity" rsc="hbtest1b"> <rule id="hbtest1b-connectivity-exclude-rule" score="-INFINITY" > <expression id="hbtest1b-connectivity-exclude" attribute="pingd-internal" operation="not_defined"/> </rule> </rsc_location> </constraints> But this causes the resource to be unrunnable on either of my both nodes, There are as expected 2 pingd daemons running: root 6132 1 0 21:07 ? 00:00:00 /usr/lib/heartbeat/pingd -D -p /var/run/heartbeat/rsctmp/pingd-pingd-internal:0 -a pingd-internal -d 5s -m 1000 -h 172.17.32.23 -h 192.168.132.23 root 13142 1 0 21:47 ? 00:00:00 /usr/lib/heartbeat/pingd -D -p /var/run/heartbeat/rsctmp/pingd-pingd-external:0 -a pingd-external -d 5s -m 1000 -h 195.244.97.241 The problem is, I can't see in the cibadmin -Q output that the pingd daemons have have stored their results anywhere.. In the log I see the following output: Sep 24 22:01:46 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=24031 Sep 24 22:01:47 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=24031 Sep 24 22:01:48 xen20a pingd: [13142]: WARN: dump_v4_echo: Bad echo (0): 8, code=0, seq=261, id=0, check=22762 Sep 24 22:01:48 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0): 8, code=0, seq=263, id=0, check=22250 Sep 24 22:01:50 xen20a pingd: [13142]: info: stand_alone_ping: Node 195.244.97.241 is alive (1) Sep 24 22:01:50 xen20a pingd: [13142]: info: send_update: 1 active ping nodes Sep 24 22:01:51 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=24031 Sep 24 22:01:52 xen20a pingd: [6132]: info: stand_alone_ping: Node 172.17.32.23 is alive (3) Sep 24 22:01:53 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=24031 Sep 24 22:01:54 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=25823 Sep 24 22:01:55 xen20a pingd: [6132]: WARN: dump_v4_echo: Bad echo (0): 3, code=1, seq=0, id=0, check=24031 Sep 24 22:01:57 xen20a pingd: [6132]: info: stand_alone_ping: Node 192.168.132.23 is alive (2) Sep 24 22:01:57 xen20a pingd: [6132]: info: send_update: 2 active ping nodes Where should the current pingd status be located in the cib ? What is wrong with my setup ? TIA, Bruno
signature.asc
Description: OpenPGP digital signature
_______________________________________________ Pacemaker mailing list Pacemaker@clusterlabs.org http://list.clusterlabs.org/mailman/listinfo/pacemaker