Hi Ken, When I ran ocf-tester to test IPaddr2 agent
*ocf-tester -n sc_vip -o ip=192.168.20.188 -o cidr_netmask=24 -o nic=eth0 /usr/lib/ocf/resource.d/heartbeat/IPaddr2* I got this error - *ERROR: Setup problem: couldn't find command: ip *in *test_command monitor. *I verified ip command is there but still this error. What might be the reason for this error ? Is this okay ? Running: export export OCF_RESOURCE_INSTANCE=sc_vip OCF_RESKEY_ip='192.168.20.188' OCF_RESKEY_cidr_netmask='24' OCF_RESKEY_nic='eth0'; bash /usr/lib/ocf/resource.d/heartbeat/IPaddr2 monitor 2>&1 > /dev/null command_output: + : /usr/lib/ocf/lib/heartbeat + . /usr/lib/ocf/lib/heartbeat/ocf-shellfuncs ++ unset LC_ALL ++ export LC_ALL ++ unset LANGUAGE ++ export LANGUAGE +++ basename /usr/lib/ocf/resource.d/heartbeat/IPaddr2 ++ __SCRIPT_NAME=IPaddr2 ++ '[' -z /usr/lib/ocf ']' ++ '[' /usr/lib/ocf/lib/heartbeat = /usr/lib/ocf/resource.d/heartbeat ']' ++ : /usr/lib/ocf/lib/heartbeat ++ . /usr/lib/ocf/lib/heartbeat/ocf-binaries +++ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/sbin:/bin:/usr/sbin:/usr/bin +++ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/sbin:/bin:/usr/sbin:/usr/bin:/usr/ucb +++ export PATH +++ : mawk +++ : /bin/grep -E +++ : +++ : mail +++ : /bin/ping +++ : /bin/bash +++ : /usr/bin/test +++ : /usr/bin/test +++ : basename +++ : blockdev +++ : cat +++ : fsck +++ : fuser +++ : getent +++ : grep +++ : ifconfig +++ : iptables +++ : ip +++ : mdadm +++ : modprobe +++ : mount +++ : msgfmt +++ : netstat +++ : perl +++ : python +++ : raidstart +++ : raidstop +++ : route +++ : umount +++ : reboot +++ : poweroff +++ : wget +++ : whoami +++ : strings +++ : scp +++ : ssh +++ : swig +++ : gzip +++ : tar +++ : md5 +++ : drbdadm +++ : drbdsetup ++ . /usr/lib/ocf/lib/heartbeat/ocf-returncodes +++ OCF_SUCCESS=0 +++ OCF_ERR_GENERIC=1 +++ OCF_ERR_ARGS=2 +++ OCF_ERR_UNIMPLEMENTED=3 +++ OCF_ERR_PERM=4 +++ OCF_ERR_INSTALLED=5 +++ OCF_ERR_CONFIGURED=6 +++ OCF_NOT_RUNNING=7 +++ OCF_RUNNING_MASTER=8 +++ OCF_FAILED_MASTER=9 ++ . /usr/lib/ocf/lib/heartbeat/ocf-directories +++ prefix=/usr +++ exec_prefix=/usr +++ : /etc/init.d +++ : /etc/ha.d +++ : /etc/ha.d/rc.d +++ : /etc/ha.d/conf +++ : /etc/ha.d/ha.cf +++ : /var/lib/heartbeat +++ : /var/run/resource-agents +++ : /var/run/heartbeat/rsctmp +++ : /var/lib/heartbeat/fifo +++ : /usr/lib/heartbeat +++ : /usr/sbin +++ : %Y/%m/%d_%T +++ : /dev/null +++ : /etc/ha.d/resource.d +++ : /usr/share/doc/heartbeat +++ : IPaddr2 +++ : /var/run/ +++ : /var/lock/subsys/ ++ . /usr/lib/ocf/lib/heartbeat/ocf-rarun ++ : 0 ++ __ocf_set_defaults monitor ++ __OCF_ACTION=monitor ++ unset LANG ++ LC_ALL=C ++ export LC_ALL ++ '[' -z '' ']' ++ : 0 ++ '[' '!' -d /usr/lib/ocf ']' ++ '[' -z '' ']' ++ : IPaddr2 ++ '[' -z '' ']' ++ : We are being invoked as an init script. ++ : Fill in some things with reasonable values. ++ : sc_vip ++ return 0 + OCF_RESKEY_lvs_support_default=false + OCF_RESKEY_clusterip_hash_default=sourceip-sourceport + OCF_RESKEY_unique_clone_address_default=false + OCF_RESKEY_arp_interval_default=200 + OCF_RESKEY_arp_count_default=5 + OCF_RESKEY_arp_bg_default=true + OCF_RESKEY_arp_mac_default=ffffffffffff + : false + : sourceip-sourceport + : false + : 200 + : 5 + : true + : ffffffffffff + SENDARP=/usr/lib/heartbeat/send_arp + FINDIF=/usr/lib/heartbeat/findif + VLDIR=/var/run/resource-agents + SENDARPPIDDIR=/var/run/resource-agents + CIP_lockfile=/var/run/resource-agents/IPaddr2-CIP-192.168.20.188 + ocf_is_true false + case "$1" in + false + case $__OCF_ACTION in + ip_validate + check_binary ip + have_binary ip + '[' 1 = 1 ']' + false + '[' 7 = 7 ']' + ocf_log err 'Setup problem: couldn'\''t find command: ip' + '[' 2 -lt 2 ']' + __OCF_PRIO=err + shift + __OCF_MSG='Setup problem: couldn'\''t find command: ip' + case "${__OCF_PRIO}" in + __OCF_PRIO=ERROR + '[' ERROR = DEBUG ']' + ha_log 'ERROR: Setup problem: couldn'\''t find command: ip' + local loglevel + '[' none = '' ']' + tty + '[' x = x0 -a x = xdebug ']' + '[' '' ']' + echo 'ERROR: Setup problem: couldn'\''t find command: ip' ERROR: Setup problem: couldn't find command: ip + return 0 + exit 5 On Thu, Oct 29, 2015 at 2:32 PM, Pritam Kharat < pritam.kha...@oneconvergence.com> wrote: > Hi Dejan > > It is giving following info. Then I tried *crm resource restart sc_vip* > too but no trace found. Anything which I need to do more apart from this ? > > root@sc-node-1:/var/lib/heartbeat# crm resource trace sc_vip stop > INFO: restart sc_vip to get the trace > > On Thu, Oct 29, 2015 at 2:10 PM, Dejan Muhamedagic <deja...@fastmail.fm> > wrote: > >> Hi, >> >> On Thu, Oct 29, 2015 at 10:40:18AM +0530, Pritam Kharat wrote: >> > Thank you very much Ken for reply. I will try your suggested steps. >> >> If you cannot figure out from the logs why the stop operation >> times out, you can also try to trace the resource agent: >> >> # crm resource help trace >> # crm resource trace vip stop >> >> Then take a look at the trace or post it somewhere. >> >> Thanks, >> >> Dejan >> >> > >> > On Wed, Oct 28, 2015 at 11:23 PM, Ken Gaillot <kgail...@redhat.com> >> wrote: >> > >> > > On 10/28/2015 03:51 AM, Pritam Kharat wrote: >> > > > Hi All, >> > > > >> > > > I am facing one issue in my two node HA. When I stop pacemaker on >> ACTIVE >> > > > node, it takes more time to stop and by this time VIP migration with >> > > other >> > > > resources migration fails to STANDBY node. (I have seen same issue >> in >> > > > ACTIVE node reboot case also) >> > > >> > > I assume STANDBY in this case is just a description of the node's >> > > purpose, and does not mean that you placed the node in pacemaker's >> > > standby mode. If the node really is in standby mode, it can't run any >> > > resources. >> > > >> > > > Last change: Wed Oct 28 02:52:57 2015 via cibadmin on node-1 >> > > > Stack: corosync >> > > > Current DC: node-1 (1) - partition with quorum >> > > > Version: 1.1.10-42f2063 >> > > > 2 Nodes configured >> > > > 2 Resources configured >> > > > >> > > > >> > > > Online: [ node-1 node-2 ] >> > > > >> > > > Full list of resources: >> > > > >> > > > resource (upstart:resource): Stopped >> > > > vip (ocf::heartbeat:IPaddr2): Started node-2 (unmanaged) FAILED >> > > > >> > > > Migration summary: >> > > > * Node node-1: >> > > > * Node node-2: >> > > > >> > > > Failed actions: >> > > > vip_stop_0 (node=node-2, call=-1, rc=1, status=Timed Out, >> > > > last-rc-change=Wed Oct 28 03:05:24 2015 >> > > > , queued=0ms, exec=0ms >> > > > ): unknown error >> > > > >> > > > VIP monitor is failing over here with error Timed Out. What is the >> > > general >> > > > reason for TimeOut. ? I have kept default-action-timeout=180secs >> which >> > > > should be enough for monitoring >> > > >> > > 180s should be far more than enough, so something must be going wrong. >> > > Notice that it is the stop operation on the active node that is >> failing. >> > > Normally in such a case, pacemaker would fence that node to be sure >> that >> > > it is safe to bring it up elsewhere, but you have disabled stonith. >> > > >> > > Fencing is important in failure recovery such as this, so it would be >> a >> > > good idea to try to get it implemented. >> > > >> > > > I have added order property -> when vip is started then only start >> other >> > > > resources. >> > > > Any clue to solve this problem ? Most of the time this VIP >> monitoring is >> > > > failing with Timed Out error. >> > > >> > > The "stop" in "vip_stop_0" means that the stop operation is what >> failed. >> > > Have you seen timeouts on any other operations? >> > > >> > > Look through the logs around the time of the failure, and try to see >> if >> > > there are any indications as to why the stop failed. >> > > >> > > If you can set aside some time for testing or have a test cluster that >> > > exhibits the same issue, you can try unmanaging the resource in >> > > pacemaker, then: >> > > >> > > 1. Try adding/removing the IP via normal system commands, and make >> sure >> > > that works. >> > > >> > > 2. Try running the resource agent manually (with any verbose option) >> to >> > > start/stop/monitor the IP to see if you can reproduce the problem and >> > > get more messages. >> > > >> > > _______________________________________________ >> > > Users mailing list: Users@clusterlabs.org >> > > http://clusterlabs.org/mailman/listinfo/users >> > > >> > > Project Home: http://www.clusterlabs.org >> > > Getting started: >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > > Bugs: http://bugs.clusterlabs.org >> > > >> > >> > >> > >> > -- >> > Thanks and Regards, >> > Pritam Kharat. >> >> > _______________________________________________ >> > Users mailing list: Users@clusterlabs.org >> > http://clusterlabs.org/mailman/listinfo/users >> > >> > Project Home: http://www.clusterlabs.org >> > Getting started: >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > Bugs: http://bugs.clusterlabs.org >> >> >> _______________________________________________ >> Users mailing list: Users@clusterlabs.org >> http://clusterlabs.org/mailman/listinfo/users >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > > > > -- > Thanks and Regards, > Pritam Kharat. > -- Thanks and Regards, Pritam Kharat.
_______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org