Hi Jitendra Yes I'm doubt that it need to enter the ssh-agent bash & ssh-add before I ssh the NN from each other.Is it an problem?
Regards 2013/12/2 Jitendra Yadav <jeetuyadav200...@gmail.com> > Are you able to connect both NN hosts using SSH without password? > Make sure you have correct ssh keys in authorized key file. > > Regards > Jitendra > > > On Mon, Dec 2, 2013 at 5:50 PM, YouPeng Yang <yypvsxf19870...@gmail.com>wrote: > >> Hi Pavan >> >> >> I'm using sshfence >> >> ------core-site.xml----------------- >> >> <configuration> >> <property> >> <name>fs.defaultFS</name> >> <value>hdfs://lklcluster</value> >> <final>true</final> >> </property> >> >> <property> >> <name>hadoop.tmp.dir</name> >> <value>/home/hadoop/tmp2</value> >> </property> >> >> >> </configuration> >> >> >> -------hdfs-site.xml------------- >> >> <configuration> >> <property> >> <name>dfs.namenode.name.dir</name> >> <value>/home/hadoop/namedir2</value> >> </property> >> >> <property> >> <name>dfs.datanode.data.dir</name> >> <value>/home/hadoop/datadir2</value> >> </property> >> >> <property> >> <name>dfs.nameservices</name> >> <value>lklcluster</value> >> </property> >> >> <property> >> <name>dfs.ha.namenodes.lklcluster</name> >> <value>nn1,nn2</value> >> </property> >> <property> >> <name>dfs.namenode.rpc-address.lklcluster.nn1</name> >> <value>hadoop2:8020</value> >> </property> >> <property> >> <name>dfs.namenode.rpc-address.lklcluster.nn2</name> >> <value>hadoop3:8020</value> >> </property> >> >> <property> >> <name>dfs.namenode.http-address.lklcluster.nn1</name> >> <value>hadoop2:50070</value> >> </property> >> >> <property> >> <name>dfs.namenode.http-address.lklcluster.nn2</name> >> <value>hadoop3:50070</value> >> </property> >> >> <property> >> <name>dfs.namenode.shared.edits.dir</name> >> >> <value>qjournal://hadoop1:8485;hadoop2:8485;hadoop3:8485/lklcluster</value> >> </property> >> <property> >> <name>dfs.client.failover.proxy.provider.lklcluster</name> >> >> <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value> >> </property> >> <property> >> <name>dfs.ha.fencing.methods</name> >> <value>sshfence</value> >> </property> >> >> <property> >> <name>dfs.ha.fencing.ssh.private-key-files</name> >> <value>/home/hadoop/.ssh/id_rsa</value> >> </property> >> >> <property> >> <name>dfs.ha.fencing.ssh.connect-timeout</name> >> <value>5000</value> >> </property> >> >> <property> >> <name>dfs.journalnode.edits.dir</name> >> <value>/home/hadoop/journal/data</value> >> </property> >> >> <property> >> <name>dfs.ha.automatic-failover.enabled</name> >> <value>true</value> >> </property> >> >> <property> >> <name>ha.zookeeper.quorum</name> >> <value>hadoop1:2181,hadoop2:2181,hadoop3:2181</value> >> </property> >> >> </configuration> >> >> >> 2013/12/2 Pavan Kumar Polineni <smartsunny...@gmail.com> >> >>> Post your config files and in which method you are following for >>> automatic failover >>> >>> >>> On Mon, Dec 2, 2013 at 5:34 PM, YouPeng Yang >>> <yypvsxf19870...@gmail.com>wrote: >>> >>>> Hi i >>>> I'm testing the HA auto-failover within hadoop-2.2.0 >>>> >>>> The cluster can be manully failover ,however failed with the >>>> automatic failover. >>>> I setup the HA according to the URL >>>> >>>> http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/HDFSHighAvailabilityWithQJM.html >>>> >>>> When I test the automatic failover, I killed my active NN by kill -9 >>>> <Pid-nn>,while the standby namenode does not change to active state. >>>> It came out the log in my DFSZKFailoverController as [1] >>>> >>>> Please help me ,any suggestion will be appreciated. >>>> >>>> >>>> Regards. >>>> >>>> >>>> zkfc >>>> log[1]---------------------------------------------------------------------------------------------------- >>>> >>>> 2013-12-02 19:49:28,588 INFO org.apache.hadoop.ha.NodeFencer: ====== >>>> Beginning Service Fencing Process... ====== >>>> 2013-12-02 19:49:28,588 INFO org.apache.hadoop.ha.NodeFencer: Trying >>>> method 1/1: org.apache.hadoop.ha.SshFenceByTcpPort(null) >>>> 2013-12-02 19:49:28,590 INFO org.apache.hadoop.ha.SshFenceByTcpPort: >>>> Connecting to hadoop3... >>>> 2013-12-02 19:49:28,590 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Connecting to hadoop3 port 22 >>>> 2013-12-02 19:49:28,592 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Connection established >>>> 2013-12-02 19:49:28,603 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Remote version string: >>>> SSH-2.0-OpenSSH_5.3 >>>> 2013-12-02 19:49:28,603 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Local version string: >>>> SSH-2.0-JSCH-0.1.42 >>>> 2013-12-02 19:49:28,603 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: CheckCiphers: >>>> aes256-ctr,aes192-ctr,aes128-ctr,aes256-cbc,aes192-cbc,aes128-cbc,3des-ctr,arcfour,arcfour128,arcfour256 >>>> 2013-12-02 19:49:28,608 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: aes256-ctr is not available. >>>> 2013-12-02 19:49:28,608 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: aes192-ctr is not available. >>>> 2013-12-02 19:49:28,608 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: aes256-cbc is not available. >>>> 2013-12-02 19:49:28,608 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: aes192-cbc is not available. >>>> 2013-12-02 19:49:28,609 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: arcfour256 is not available. >>>> 2013-12-02 19:49:28,610 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_KEXINIT sent >>>> 2013-12-02 19:49:28,610 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_KEXINIT received >>>> 2013-12-02 19:49:28,610 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: kex: server->client aes128-ctr >>>> hmac-md5 none >>>> 2013-12-02 19:49:28,610 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: kex: client->server aes128-ctr >>>> hmac-md5 none >>>> 2013-12-02 19:49:28,617 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_KEXDH_INIT sent >>>> 2013-12-02 19:49:28,617 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: expecting SSH_MSG_KEXDH_REPLY >>>> 2013-12-02 19:49:28,634 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: ssh_rsa_verify: signature true >>>> 2013-12-02 19:49:28,635 WARN >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Permanently added 'hadoop3' >>>> (RSA) to the list of known hosts. >>>> 2013-12-02 19:49:28,635 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_NEWKEYS sent >>>> 2013-12-02 19:49:28,635 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_NEWKEYS received >>>> 2013-12-02 19:49:28,636 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_SERVICE_REQUEST sent >>>> 2013-12-02 19:49:28,637 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: SSH_MSG_SERVICE_ACCEPT >>>> received >>>> 2013-12-02 19:49:28,638 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Authentications that can >>>> continue: gssapi-with-mic,publickey,keyboard-interactive,password >>>> 2013-12-02 19:49:28,639 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Next authentication method: >>>> gssapi-with-mic >>>> 2013-12-02 19:49:28,642 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Authentications that can >>>> continue: publickey,keyboard-interactive,password >>>> 2013-12-02 19:49:28,642 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Next authentication method: >>>> publickey >>>> 2013-12-02 19:49:28,644 INFO >>>> org.apache.hadoop.ha.SshFenceByTcpPort.jsch: Disconnecting from hadoop3 >>>> port 22 >>>> 2013-12-02 19:49:28,644 WARN org.apache.hadoop.ha.SshFenceByTcpPort: >>>> Unable to connect to hadoop3 as user hadoop >>>> com.jcraft.jsch.JSchException: Auth fail >>>> at com.jcraft.jsch.Session.connect(Session.java:452) >>>> at >>>> org.apache.hadoop.ha.SshFenceByTcpPort.tryFence(SshFenceByTcpPort.java:100) >>>> at org.apache.hadoop.ha.NodeFencer.fence(NodeFencer.java:97) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.doFence(ZKFailoverController.java:521) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.fenceOldActive(ZKFailoverController.java:494) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.access$1100(ZKFailoverController.java:59) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController$ElectorCallbacks.fenceOldActive(ZKFailoverController.java:837) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.fenceOldActive(ActiveStandbyElector.java:900) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.becomeActive(ActiveStandbyElector.java:799) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.processResult(ActiveStandbyElector.java:415) >>>> at >>>> org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:596) >>>> at >>>> org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495) >>>> 2013-12-02 19:49:28,644 WARN org.apache.hadoop.ha.NodeFencer: Fencing >>>> method org.apache.hadoop.ha.SshFenceByTcpPort(null) was unsuccessful. >>>> 2013-12-02 19:49:28,645 ERROR org.apache.hadoop.ha.NodeFencer: Unable >>>> to fence service by any configured method. >>>> 2013-12-02 19:49:28,645 INFO org.apache.hadoop.ha.ZKFailoverController: >>>> Local service NameNode at hadoop2/10.7.23.125:8020 entered state: >>>> SERVICE_NOT_RESPONDING >>>> 2013-12-02 19:49:28,646 WARN org.apache.hadoop.ha.ActiveStandbyElector: >>>> Exception handling the winning of election >>>> java.lang.RuntimeException: Unable to fence NameNode at hadoop3/ >>>> 10.7.23.124:8020 >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.doFence(ZKFailoverController.java:522) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.fenceOldActive(ZKFailoverController.java:494) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController.access$1100(ZKFailoverController.java:59) >>>> at >>>> org.apache.hadoop.ha.ZKFailoverController$ElectorCallbacks.fenceOldActive(ZKFailoverController.java:837) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.fenceOldActive(ActiveStandbyElector.java:900) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.becomeActive(ActiveStandbyElector.java:799) >>>> at >>>> org.apache.hadoop.ha.ActiveStandbyElector.processResult(ActiveStandbyElector.java:415) >>>> at >>>> org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:596) >>>> at >>>> org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495) >>>> 2013-12-02 19:49:28,646 INFO org.apache.hadoop.ha.ActiveStandbyElector: >>>> Trying to re-establish ZK session >>>> 2013-12-02 19:49:28,669 INFO org.apache.zookeeper.ZooKeeper: Session: >>>> 0x2429313c808025b closed >>>> 2013-12-02 19:49:29,672 INFO org.apache.zookeeper.ZooKeeper: Initiating >>>> client connection, connectString=hadoop1:2181,hadoop2:2181,hadoop3:2181 >>>> sessionTimeout=5000 >>>> watcher=org.apache.hadoop.ha.ActiveStandbyElector$WatcherWithClientRef@3545fe3b >>>> 2013-12-02 19:49:29,675 INFO org.apache.zookeeper.ClientCnxn: Opening >>>> socket connection to server hadoop3/10.7.23.124:2181. Will not attempt >>>> to authenticate using SASL (Unable to locate a login configuration) >>>> 2013-12-02 19:49:29,675 INFO org.apache.zookeeper.ClientCnxn: Socket >>>> connection established to hadoop3/10.7.23.124:2181, initiating session >>>> 2013-12-02 19:49:29,699 INFO org.apache.zookeeper.ClientCnxn: Session >>>> establishment complete on server hadoop3/10.7.23.124:2181, sessionid = >>>> 0x3429312ba330262, negotiated timeout = 5000 >>>> 2013-12-02 19:49:29,702 INFO org.apache.zookeeper.ClientCnxn: >>>> EventThread shut down >>>> 2013-12-02 19:49:29,706 INFO org.apache.hadoop.ha.ActiveStandbyElector: >>>> Session connected. >>>> 2013-12-02 19:49:29,706 INFO org.apache.hadoop.ha.ZKFailoverController: >>>> Quitting master election for NameNode at hadoop2/10.7.23.125:8020 and >>>> marking that fencing is necessary >>>> 2013-12-02 19:49:29,706 INFO org.apache.hadoop.ha.ActiveStandbyElector: >>>> Yielding from election >>>> 2013-12-02 19:49:29,727 INFO org.apache.zookeeper.ZooKeeper: Session: >>>> 0x3429312ba330262 closed >>>> 2013-12-02 19:49:29,728 WARN org.apache.hadoop.ha.ActiveStandbyElector: >>>> Ignoring stale result from old client with sessionId 0x3429312ba330262 >>>> 2013-12-02 19:49:29,728 INFO org.apache.zookeeper.ClientCnxn: >>>> EventThread shut down >>>> >>> >>> >> >