ncomplete=10, Source=/var/lib/pacemaker/pengine/pe-warn-7.bz2): Stopped
Jul 03 14:10:51 [2701] sip2       crmd:   notice:
too_many_st_failures:         No devices found in cluster to fence
sip1, giving up

Jul 03 14:10:54 [2697] sip2 stonith-ng:     info: stonith_command:
 Processed st_query reply from sip2: OK (0)
Jul 03 14:10:54 [2697] sip2 stonith-ng:    error: remote_op_done:
 Operation reboot of sip1 by sip2 for
stonith_admin.cman.28299@sip2.94474607: No such device

Jul 03 14:10:54 [2697] sip2 stonith-ng:     info: stonith_command:
 Processed st_notify reply from sip2: OK (0)
Jul 03 14:10:54 [2701] sip2       crmd:   notice:
tengine_stonith_notify:       Peer sip1 was not terminated (reboot) by
sip2 for sip2: No such device
(ref=94474607-8cd2-410d-bbf7-5bc7df614a50) by client
stonith_admin.cman.28299

:::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::

Sorry for the short answer, have you tested your cluster fencing ? can
you show your cluster.conf xml?

2014-08-14 14:44 GMT+02:00 Miha <m...@softnet.si>:
> emmanuel,
>
> tnx. But how to know why fancing stop working?
>
> br
> miha
>
> Dne 8/14/2014 2:35 PM, piše emmanuel segura:
>
>> Node sip2: UNCLEAN (offline) is unclean because the cluster fencing
>> failed to complete the operation
>>
>> 2014-08-14 14:13 GMT+02:00 Miha <m...@softnet.si>:
>>>
>>> hi.
>>>
>>> another thing.
>>>
>>> On node I pcs is running:
>>> [root@sip1 ~]# pcs status
>>> Cluster name: sipproxy
>>> Last updated: Thu Aug 14 14:13:37 2014
>>> Last change: Sat Feb  1 20:10:48 2014 via crm_attribute on sip1
>>> Stack: cman
>>> Current DC: sip1 - partition with quorum
>>> Version: 1.1.10-14.el6-368c726
>>> 2 Nodes configured
>>> 10 Resources configured
>>>
>>>
>>> Node sip2: UNCLEAN (offline)
>>> Online: [ sip1 ]
>>>
>>> Full list of resources:
>>>
>>>   Master/Slave Set: ms_drbd_mysql [p_drbd_mysql]
>>>       Masters: [ sip2 ]
>>>       Slaves: [ sip1 ]
>>>   Resource Group: g_mysql
>>>       p_fs_mysql (ocf::heartbeat:Filesystem):    Started sip2
>>>       p_ip_mysql (ocf::heartbeat:IPaddr2):       Started sip2
>>>       p_mysql    (ocf::heartbeat:mysql): Started sip2
>>>   Clone Set: cl_ping [p_ping]
>>>       Started: [ sip1 sip2 ]
>>>   opensips       (lsb:opensips): Stopped
>>>   fence_sip1     (stonith:fence_bladecenter_snmp):       Started sip2
>>>   fence_sip2     (stonith:fence_bladecenter_snmp):       Started sip2
>>>
>>>
>>> [root@sip1 ~]#
>>>
>>>
>>>
>>>
>>>
>>> Dne 8/14/2014 2:12 PM, piše Miha:
>>>
>>>> Hi emmanuel,
>>>>
>>>> i think so, what is the best way to check?
>>>>
>>>> Sorry for my noob question, I have confiured this 6 mouths ago and
>>>> everything was working fine till now. Now I need to find out what realy
>>>> heppend beffor I do something stupid.
>>>>
>>>>
>>>>
>>>> tnx
>>>>
>>>> Dne 8/14/2014 1:58 PM, piše emmanuel segura:
>>>>>
>>>>> are you sure your cluster fencing is working?
>>>>>
>>>>> 2014-08-14 13:40 GMT+02:00 Miha <m...@softnet.si>:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I noticed today that I am having some problem with cluster. I noticed
>>>>>> the
>>>>>> master server is offilne but still virutal ip is assigned to it and
>>>>>> all
>>>>>> services are running properly (for production).
>>>>>>
>>>>>> If I do this I am getting this notifications:
>>>>>>
>>>>>> [root@sip2 cluster]# pcs status
>>>>>> Error: cluster is not currently running on this node
>>>>>> [root@sip2 cluster]# /etc/init.d/corosync status
>>>>>> corosync dead but pid file exists
>>>>>> [root@sip2 cluster]# pcs status
>>>>>> Error: cluster is not currently running on this node
>>>>>> [root@sip2 cluster]#
>>>>>> [root@sip2 cluster]#
>>>>>> [root@sip2 cluster]# tailf fenced.log
>>>>>> Aug 14 13:34:25 fenced cman_get_cluster error -1 112
>>>>>>
>>>>>>
>>>>>> The main thing is what to do now? Do "pcs start" and hope for the best
>>>>>> or
>>>>>> what?
>>>>>>
>>>>>> I have pasted log in pastebin: http://pastebin.com/SUp2GcmN
>>>>>>
>>>>>> tnx!
>>>>>>
>>>>>> miha
>>>>>>
>>>>>> _______________________________________________
>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>>
>>>>>
>>>
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>
>>
>>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org



-- 
esta es mi vida e me la vivo hasta que dios quiera

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to