I installed the hadoop by untar the hadoop-2.6.0.tar.gz, will check further. Thanks.
2014-12-16 14:39 GMT-06:00 Jiayu Ji <jiayu...@gmail.com>: > > The cluster is running hadoop 2.0 while you client side in under hadoop > 1.0. > > I would guess you have installed 1.0 on your client machine before and > your env variable is still pointing to it. > > On Tue, Dec 16, 2014 at 9:31 AM, Dan Dong <dongda...@gmail.com> wrote: >> >> Thanks, the error now changes to the following: >> $ hadoop dfsadmin -report >> report: Server IPC version 9 cannot communicate with client version 4 >> >> Not clear which Server and which client are conflicting. All hadoop >> components comes from the hadoop-2.6.0.tar.gz package, what's going wrong? >> >> Cheers, >> Dan >> >> >> 2014-12-15 22:30 GMT-06:00 Susheel Kumar Gadalay <skgada...@gmail.com>: >> >>> Give complete hostname with domain name not just master-node. >>> >>> <property> >>> <name>fs.defaultFS</name> >>> <value>hdfs://master-node.domain.name:9000</value> >>> </property> >>> >>> Else give IP address also >>> >>> >>> On 12/16/14, Dan Dong <dongda...@gmail.com> wrote: >>> > Hi, Johny, >>> > Yes, they have been turned off from the beginning. Guess the problem >>> is >>> > still in the conf files, it would be helpful if some example *.xml >>> could be >>> > shown. >>> > >>> > Cheers, >>> > Dan >>> > >>> > >>> > 2014-12-15 12:24 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >>> >> >>> >> do you have selinux and iptables turned off? >>> >> >>> >> ------------------------------ >>> >> Date: Mon, 15 Dec 2014 09:54:41 -0600 >>> >> Subject: Re: Hadoop 2.6.0: "FileSystem file:/// is not a distributed >>> file >>> >> system" >>> >> From: dongda...@gmail.com >>> >> To: user@hadoop.apache.org >>> >> >>> >> >>> >> Found in the log file: >>> >> 2014-12-12 15:51:10,434 ERROR >>> >> org.apache.hadoop.hdfs.server.namenode.NameNode: >>> >> java.lang.IllegalArgumentException: Does not contain a valid host:port >>> >> authority: file:/// >>> >> at >>> >> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:212) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:280) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:569) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1479) >>> >> at >>> >> >>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1488) >>> >> >>> >> But I have set it in core-site.xml already: >>> >> <property> >>> >> <name>fs.defaultFS</name> >>> >> <value>hdfs://master-node:9000</value> >>> >> </property> >>> >> >>> >> Other settings: >>> >> $ cat mapred-site.xml >>> >> <configuration> >>> >> <property> >>> >> <name>mapred.job.tracker</name> >>> >> <value>master-node:9002</value> >>> >> </property> >>> >> <property> >>> >> <name>mapreduce.jobhistory.address</name> >>> >> <value>master-node:10020</value> >>> >> </property> >>> >> <property> >>> >> <name>mapreduce.jobhistory.webapp.address</name> >>> >> <value>master-node:19888</value> >>> >> </property> >>> >> </configuration> >>> >> >>> >> $ cat yarn-site.xml >>> >> <configuration> >>> >> >>> >> <!-- Site specific YARN configuration properties --> >>> >> <property> >>> >> <name>mapreduce.framework.name</name> >>> >> <value>yarn</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.resourcemanager.address</name> >>> >> <value>master-node:18040</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.resourcemanager.scheduler.address</name> >>> >> <value>master-node:18030</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.resourcemanager.webapp.address</name> >>> >> <value>master-node:18088</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.resourcemanager.resource-tracker.address</name> >>> >> <value>master-node:18025</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.resourcemanager.admin.address</name> >>> >> <value>master-node:18141</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.nodemanager.aux-services</name> >>> >> <value>mapreduce_shuffle</value> >>> >> </property> >>> >> <property> >>> >> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> >>> >> <value>org.apache.hadoop.mapred.ShuffleHandler</value> >>> >> </property> >>> >> </configuration> >>> >> >>> >> Cheers, >>> >> Dan >>> >> >>> >> >>> >> 2014-12-15 9:17 GMT-06:00 Dan Dong <dongda...@gmail.com>: >>> >> >>> >> Thank you all, but still the same after change file:/ to file://, and >>> >> HADOOP_CONF_DIR points to the correct position already: >>> >> $ echo $HADOOP_CONF_DIR >>> >> /home/dong/import/hadoop-2.6.0/etc/hadoop >>> >> >>> >> >>> >> 2014-12-15 8:57 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >>> >> >>> >> Don't you have to use file:// instead of just one /? >>> >> >>> >> ------------------------------ >>> >> From: brahmareddy.batt...@huawei.com >>> >> To: user@hadoop.apache.org >>> >> Subject: RE: Hadoop 2.6.0: "FileSystem file:/// is not a distributed >>> file >>> >> system" >>> >> Date: Sat, 13 Dec 2014 05:48:18 +0000 >>> >> >>> >> >>> >> Hi Dong, >>> >> >>> >> HADOOP_CONF_DIR might be referring to default..you can export >>> >> HADOOP_CONF_DIR where following configuration files are present.. >>> >> >>> >> Thanks & Regards >>> >> Brahma Reddy Battula >>> >> >>> >> >>> >> ------------------------------ >>> >> *From:* Dan Dong [dongda...@gmail.com] >>> >> *Sent:* Saturday, December 13, 2014 3:43 AM >>> >> *To:* user@hadoop.apache.org >>> >> *Subject:* Hadoop 2.6.0: "FileSystem file:/// is not a distributed >>> file >>> >> system" >>> >> >>> >> Hi, >>> >> I installed Hadoop2.6.0 on my cluster with 2 nodes, I got the >>> following >>> >> error when I run: >>> >> $hadoop dfsadmin -report >>> >> FileSystem file:/// is not a distributed file system >>> >> >>> >> What this mean? I have set it in core-site.xml already: >>> >> <property> >>> >> <name>fs.defaultFS</name> >>> >> <value>hdfs://master-node:9000</value> >>> >> </property> >>> >> >>> >> and in hdfs-site.xml: >>> >> <property> >>> >> <name>dfs.namenode.name.dir</name> >>> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/name</value> >>> >> <final>true</final> >>> >> </property> >>> >> <property> >>> >> <name>dfs.dataname.data.dir</name> >>> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/data</value> >>> >> <final>true</final> >>> >> </property> >>> >> >>> >> The java process are running on master as: >>> >> 10479 SecondaryNameNode >>> >> 10281 NameNode >>> >> 10628 ResourceManager >>> >> >>> >> and on slave: >>> >> 22870 DataNode >>> >> 22991 NodeManager >>> >> >>> >> Any hints? Thanks! >>> >> >>> >> Cheers, >>> >> Dan >>> >> >>> >> >>> >> >>> > >>> >> > > -- > Regards, > > Jiayu (James) Ji, > > Cell: (312)823-7393 > >