On 14 Jul 2014, at 10:09 pm, Dvorak Andreas <andreas.dvo...@baaderbank.de> 
wrote:

> Hi,
> 
> thank you for you answer, here is the creatation of the fancing resources:
> pcs stonith create ipmi-fencing-sv2827 fence_ipmilan pcmk_host_list="sv2827" 
> ipaddr=10.110.28.27 action="off" login=ipmi passwd=XXXXX delay=15 op monitor 
> interval=60s

So you have pcmk_host_list="sv2827" but the cluster thinks your node name is 
sv2827-p1

> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
> can_fence_host_with_device: ipmi-fencing-sv2827 can not fence sv2827-p1: 
> static-list



> pcs stonith create ipmi-fencing-sv2828 fence_ipmilan pcmk_host_list="sv2828 " 
> ipaddr=10.110.28.28 action="off" login=ipmi passwd=XXXXX delay=15 op monitor 
> interval=60s
> pcs constraint location ipmi-fencing-sv2827 prefers sv2828=-INFINITY
> pcs constraint location ipmi-fencing-sv2828 prefers sv2827=-INFINITY
> pcs property set stonith-enabled=true
> pcs property set no-quorum-policy=ignore
> 
> Best regards
> Andreas
> 
> -----Ursprüngliche Nachricht-----
> Von: Andrew Beekhof [mailto:and...@beekhof.net] 
> Gesendet: Freitag, 11. Juli 2014 01:42
> An: The Pacemaker cluster resource manager
> Betreff: Re: [Pacemaker] pacemaker stonith No such device
> 
> 
> On 9 Jul 2014, at 8:53 pm, Dvorak Andreas <andreas.dvo...@baaderbank.de> 
> wrote:
> 
>> Dear all,
>> 
>> unfortunately my stonith does not work on my pacemaker cluster. If I do 
>> ifdown on the two cluster interconnect interfaces of server sv2827 the 
>> server sv2828 want to fence the server sv2827, but the messages log says:    
>> error: remote_op_done: Operation reboot of sv2827-p1 by sv2828-p1 for 
>> crmd.7979@sv2828-p1.076062f0: No such device
>> Can somebody please help me?
>> 
> 
> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
> can_fence_host_with_device: ipmi-fencing-sv2827 can not fence sv2827-p1: 
> static-list
> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
> can_fence_host_with_device: ipmi-fencing-sv2828 can not fence sv2827-p1: 
> static-list
> 
> what does the config of your two fencing resources look like?
> 
> 
>> 
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [CMAN  ] quorum lost, blocking 
>> activity
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [QUORUM] This node is within the 
>> non-primary component and will NOT provide any services.
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [QUORUM] Members[1]: 1
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [TOTEM ] A processor joined or left 
>> the membership and a new membership was formed.
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: cman_event_callback: Membership 
>> 1492: quorum lost
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: crm_update_peer_state: 
>> cman_event_callback: Node sv2827-p1[2] - state is now lost (was member)
>> Jul  9 12:42:49 sv2828 crmd[7979]:  warning: match_down_event: No match for 
>> shutdown action on sv2827-p1
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: peer_update_callback: 
>> Stonith/shutdown of sv2827-p1 not matched
>> Jul  9 12:42:49 sv2828 kernel: dlm: closing connection to node 2
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [CPG   ] chosen downlist: sender 
>> r(0) ip(192.168.2.28) r(1) ip(192.168.3.28) ; members(old:2 left:1)
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: do_state_transition: State 
>> transition S_IDLE -> S_INTEGRATION [ input=I_NODE_JOIN cause=C_FSA_INTERNAL 
>> origin=check_join_state ]
>> Jul  9 12:42:49 sv2828 corosync[7749]:   [MAIN  ] Completed service 
>> synchronization, ready to provide service.
>> Jul  9 12:42:49 sv2828 crmd[7979]:  warning: match_down_event: No match for 
>> shutdown action on sv2827-p1
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: peer_update_callback: 
>> Stonith/shutdown of sv2827-p1 not matched
>> Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_local_callback: Sending 
>> full refresh (origin=crmd)
>> Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
>> flush op to all hosts for: last-failure-MYSQLFS (1404902183)
>> Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
>> flush op to all hosts for: probe_complete (true)
>> Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
>> flush op to all hosts for: last-failure-MYSQL (1404901921)
>> Jul  9 12:42:49 sv2828 pengine[7978]:   notice: unpack_config: On loss of 
>> CCM Quorum: Ignore
>> Jul  9 12:42:49 sv2828 pengine[7978]:  warning: pe_fence_node: Node 
>> sv2827-p1 will be fenced because the node is no longer part of the 
>> cluster Jul  9 12:42:49 sv2828 pengine[7978]:  warning: 
>> determine_online_status: Node sv2827-p1 is unclean Jul  9 12:42:49 sv2828 
>> pengine[7978]:  warning: custom_action: Action ipmi-fencing-sv2828_stop_0 on 
>> sv2827-p1 is unrunnable (offline) Jul  9 12:42:49 sv2828 pengine[7978]:  
>> warning: stage6: Scheduling Node sv2827-p1 for STONITH
>> Jul  9 12:42:49 sv2828 pengine[7978]:   notice: LogActions: Move    
>> ipmi-fencing-sv2828#011(Started sv2827-p1 -> sv2828-p1)
>> Jul  9 12:42:49 sv2828 pengine[7978]:  warning: process_pe_message: 
>> Calculated Transition 38: /var/lib/pacemaker/pengine/pe-warn-28.bz2
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: te_fence_node: Executing reboot 
>> fencing operation (20) on sv2827-p1 (timeout=60000)
>> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: handle_request: Client 
>> crmd.7979.6c35e3f1 wants to fence (reboot) 'sv2827-p1' with device '(any)'
>> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
>> initiate_remote_stonith_op: Initiating remote operation reboot for 
>> sv2827-p1: 076062f0-eff3-4798-a504-16c5c5666a5b (0)
>> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
>> can_fence_host_with_device: ipmi-fencing-sv2827 can not fence sv2827-p1: 
>> static-list
>> Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: 
>> can_fence_host_with_device: ipmi-fencing-sv2828 can not fence sv2827-p1: 
>> static-list
>> Jul  9 12:42:49 sv2828 stonith-ng[7975]:    error: remote_op_done: Operation 
>> reboot of sv2827-p1 by sv2828-p1 for crmd.7979@sv2828-p1.076062f0: No such 
>> device
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: tengine_stonith_callback: 
>> Stonith operation 8/20:38:0:71703806-8a7c-447f-a033-e3c26abd607c: No such 
>> device (-19)
>> Jul  9 12:42:49 sv2828 crmd[7979]:   notice: run_graph: Transition 38 
>> (Complete=1, Pending=0, Fired=0, Skipped=5, Incomplete=0, 
>> Source=/var/lib/pacemaker/pengine/pe-warn-28.bz2): Stopped
>> 
>> With the ipmitool I could test the correct work of the ipmi like power cycle.
>> 
>> pcs status
>> Cluster name: mysql-int-prod
>> Last updated: Wed Jul  9 12:46:43 2014 Last change: Wed Jul  9 
>> 12:41:14 2014 via crm_resource on sv2828-p1
>> Stack: cman
>> Current DC: sv2828-p1 - partition with quorum
>> Version: 1.1.10-1.el6_4.4-368c726
>> 2 Nodes configured
>> 5 Resources configured
>> Online: [ sv2827-p1 sv2828-p1 ]
>> Full list of resources:
>> ipmi-fencing-sv2827    (stonith:fence_ipmilan):            Started sv2828-p1
>> ipmi-fencing-sv2828    (stonith:fence_ipmilan):            Started sv2827-p1
>> ClusterIP            (ocf::heartbeat:IPaddr2):            Started sv2828-p1
>> MYSQLFS            (ocf::baader:MYSQLFS):               Started sv2828-p1
>> MYSQL (ocf::baader:MYSQL):    Started sv2828-p1
>> 
>> 
>> pcs property
>> Cluster Properties:
>> cluster-infrastructure: cman
>> dc-version: 1.1.10-1.el6_4.4-368c726
>> last-lrm-refresh: 1404902348
>> no-quorum-policy: ignore
>> stonith-enabled: true
>> 
>> 
>> Best regards
>> Andreas Dvorak
>> 
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org 
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> 
>> Project Home: http://www.clusterlabs.org Getting started: 
>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to