Please use jdk 6 latest. best, Alex
-- Alexander Lorenz http://mapredit.blogspot.com On Feb 9, 2012, at 11:11 AM, hadoop hive wrote: > did you make check the ssh between localhost means its should be ssh password > less between localhost > > public-key =authorized_key > > On Thu, Feb 9, 2012 at 1:06 AM, Robin Mueller-Bady > <robin.mueller-b...@oracle.com> wrote: > Dear Guruprasad, > > it would be very helpful to provide details from your configuration files as > well as more details on your setup. > It seems to be that the connection from slave to master cannot be established > ("Connection reset by peer"). > Do you use a virtual environment, physical master/slaves or all on one > machine ? > Please paste also the output of "kingul2" namenode logs. > > Regards, > > Robin > > > On 02/08/12 13:06, Guruprasad B wrote: >> Hi, >> >> I am Guruprasad from Bangalore (India). I need help in setting up hadoop >> platform. I am very much new to Hadoop Platform. >> >> I am following the below given articles and I was able to set up >> "Single-Node Cluster >> " >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/#what-we-want-to-do >> >> Now I am trying to set up " >> Multi-Node Cluster" by following the below given >> article. >> >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node-cluster/ >> >> >> Below given is my setup: >> Hadoop : hadoop_0.20.2 >> Linux: Ubuntu Linux 10.10 >> Java: java-7-oracle >> >> >> I have successfully reached till the topic "Starting the multi-node >> cluster" in the above given article. >> When I start the HDFS/MapReduce daemons it is getting started and going >> down immediately both in master & slave as well, >> please have a look at the below logs, >> >> hduser@kinigul2:/usr/local/hadoop$ bin/start-dfs.sh >> starting namenode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hduser-namenode-kinigul2.out >> master: starting datanode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hduser-datanode-kinigul2.out >> slave: starting datanode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hduser-datanode-guruL.out >> master: starting secondarynamenode, logging to >> /usr/local/hadoop/bin/../logs/hadoop-hduser-secondarynamenode-kinigul2.out >> >> hduser@kinigul2:/usr/local/hadoop$ jps >> 6098 DataNode >> 6328 Jps >> 5914 NameNode >> 6276 SecondaryNameNode >> >> hduser@kinigul2:/usr/local/hadoop$ jps >> 6350 Jps >> >> >> I am getting below given error in slave logs: >> >> 2012-02-08 21:04:01,641 ERROR >> org.apache.hadoop.hdfs.server.datanode.DataNode: java.io.IOException: Call >> to master/ >> 16.150.98.62:54310 >> failed on local exception: >> java.io.IOException: Connection reset by peer >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:775) >> at org.apache.hadoop.ipc.Client.call(Client.java:743) >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >> at $Proxy4.getProtocolVersion(Unknown Source) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:346) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:383) >> at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:314) >> at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:291) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:269) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:216) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1283) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1238) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1246) >> at >> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1368) >> Caused by: java.io.IOException: Connection reset by peer >> at sun.nio.ch.FileDispatcherImpl.read0(Native Method) >> at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39) >> at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:218) >> at sun.nio.ch.IOUtil.read(IOUtil.java:191) >> at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:359) >> at >> org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155) >> at >> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128) >> at java.io.FilterInputStream.read(FilterInputStream.java:133) >> at >> org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:276) >> at java.io.BufferedInputStream.fill(BufferedInputStream.java:235) >> at java.io.BufferedInputStream.read(BufferedInputStream.java:254) >> at java.io.DataInputStream.readInt(DataInputStream.java:387) >> at >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501) >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446) >> >> >> Can you please tell what could be the reason behind this or point me to >> some pointers? >> >> Regards, >> Guruprasad >> >> > > -- > > Robin Müller-Bady | Sales Consultant > Phone: +49 211 74839 701 | Mobile: +49 172 8438346 > Oracle STCC Fusion Middleware > > ORACLE Deutschland B.V. & Co. KG | Hamborner Strasse 51 | 40472 Düsseldorf > > ORACLE Deutschland B.V. & Co. KG > Hauptverwaltung: Riesstr. 25, D-80992 München > Registergericht: Amtsgericht München, HRA 95603 > Geschäftsführer: Jürgen Kunz > > Komplementärin: ORACLE Deutschland Verwaltung B.V. > Hertogswetering 163/167, 3543 AS Utrecht, Niederlande > Handelsregister der Handelskammer Midden-Niederlande, Nr. 30143697 > Geschäftsführer: Alexander van der Ven, Astrid Kepper, Val Maher > > Oracle is committed to developing practices and products that help > protect the environment >