The cluster is running hadoop 2.0 while you client side in under hadoop 1.0.
I would guess you have installed 1.0 on your client machine before and your env variable is still pointing to it. On Tue, Dec 16, 2014 at 9:31 AM, Dan Dong <dongda...@gmail.com> wrote: > > Thanks, the error now changes to the following: > $ hadoop dfsadmin -report > report: Server IPC version 9 cannot communicate with client version 4 > > Not clear which Server and which client are conflicting. All hadoop > components comes from the hadoop-2.6.0.tar.gz package, what's going wrong? > > Cheers, > Dan > > > 2014-12-15 22:30 GMT-06:00 Susheel Kumar Gadalay <skgada...@gmail.com>: > >> Give complete hostname with domain name not just master-node. >> >> <property> >> <name>fs.defaultFS</name> >> <value>hdfs://master-node.domain.name:9000</value> >> </property> >> >> Else give IP address also >> >> >> On 12/16/14, Dan Dong <dongda...@gmail.com> wrote: >> > Hi, Johny, >> > Yes, they have been turned off from the beginning. Guess the problem >> is >> > still in the conf files, it would be helpful if some example *.xml >> could be >> > shown. >> > >> > Cheers, >> > Dan >> > >> > >> > 2014-12-15 12:24 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >> >> >> >> do you have selinux and iptables turned off? >> >> >> >> ------------------------------ >> >> Date: Mon, 15 Dec 2014 09:54:41 -0600 >> >> Subject: Re: Hadoop 2.6.0: "FileSystem file:/// is not a distributed >> file >> >> system" >> >> From: dongda...@gmail.com >> >> To: user@hadoop.apache.org >> >> >> >> >> >> Found in the log file: >> >> 2014-12-12 15:51:10,434 ERROR >> >> org.apache.hadoop.hdfs.server.namenode.NameNode: >> >> java.lang.IllegalArgumentException: Does not contain a valid host:port >> >> authority: file:/// >> >> at >> >> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:212) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:280) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:569) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1479) >> >> at >> >> >> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1488) >> >> >> >> But I have set it in core-site.xml already: >> >> <property> >> >> <name>fs.defaultFS</name> >> >> <value>hdfs://master-node:9000</value> >> >> </property> >> >> >> >> Other settings: >> >> $ cat mapred-site.xml >> >> <configuration> >> >> <property> >> >> <name>mapred.job.tracker</name> >> >> <value>master-node:9002</value> >> >> </property> >> >> <property> >> >> <name>mapreduce.jobhistory.address</name> >> >> <value>master-node:10020</value> >> >> </property> >> >> <property> >> >> <name>mapreduce.jobhistory.webapp.address</name> >> >> <value>master-node:19888</value> >> >> </property> >> >> </configuration> >> >> >> >> $ cat yarn-site.xml >> >> <configuration> >> >> >> >> <!-- Site specific YARN configuration properties --> >> >> <property> >> >> <name>mapreduce.framework.name</name> >> >> <value>yarn</value> >> >> </property> >> >> <property> >> >> <name>yarn.resourcemanager.address</name> >> >> <value>master-node:18040</value> >> >> </property> >> >> <property> >> >> <name>yarn.resourcemanager.scheduler.address</name> >> >> <value>master-node:18030</value> >> >> </property> >> >> <property> >> >> <name>yarn.resourcemanager.webapp.address</name> >> >> <value>master-node:18088</value> >> >> </property> >> >> <property> >> >> <name>yarn.resourcemanager.resource-tracker.address</name> >> >> <value>master-node:18025</value> >> >> </property> >> >> <property> >> >> <name>yarn.resourcemanager.admin.address</name> >> >> <value>master-node:18141</value> >> >> </property> >> >> <property> >> >> <name>yarn.nodemanager.aux-services</name> >> >> <value>mapreduce_shuffle</value> >> >> </property> >> >> <property> >> >> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> >> >> <value>org.apache.hadoop.mapred.ShuffleHandler</value> >> >> </property> >> >> </configuration> >> >> >> >> Cheers, >> >> Dan >> >> >> >> >> >> 2014-12-15 9:17 GMT-06:00 Dan Dong <dongda...@gmail.com>: >> >> >> >> Thank you all, but still the same after change file:/ to file://, and >> >> HADOOP_CONF_DIR points to the correct position already: >> >> $ echo $HADOOP_CONF_DIR >> >> /home/dong/import/hadoop-2.6.0/etc/hadoop >> >> >> >> >> >> 2014-12-15 8:57 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >> >> >> >> Don't you have to use file:// instead of just one /? >> >> >> >> ------------------------------ >> >> From: brahmareddy.batt...@huawei.com >> >> To: user@hadoop.apache.org >> >> Subject: RE: Hadoop 2.6.0: "FileSystem file:/// is not a distributed >> file >> >> system" >> >> Date: Sat, 13 Dec 2014 05:48:18 +0000 >> >> >> >> >> >> Hi Dong, >> >> >> >> HADOOP_CONF_DIR might be referring to default..you can export >> >> HADOOP_CONF_DIR where following configuration files are present.. >> >> >> >> Thanks & Regards >> >> Brahma Reddy Battula >> >> >> >> >> >> ------------------------------ >> >> *From:* Dan Dong [dongda...@gmail.com] >> >> *Sent:* Saturday, December 13, 2014 3:43 AM >> >> *To:* user@hadoop.apache.org >> >> *Subject:* Hadoop 2.6.0: "FileSystem file:/// is not a distributed file >> >> system" >> >> >> >> Hi, >> >> I installed Hadoop2.6.0 on my cluster with 2 nodes, I got the >> following >> >> error when I run: >> >> $hadoop dfsadmin -report >> >> FileSystem file:/// is not a distributed file system >> >> >> >> What this mean? I have set it in core-site.xml already: >> >> <property> >> >> <name>fs.defaultFS</name> >> >> <value>hdfs://master-node:9000</value> >> >> </property> >> >> >> >> and in hdfs-site.xml: >> >> <property> >> >> <name>dfs.namenode.name.dir</name> >> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/name</value> >> >> <final>true</final> >> >> </property> >> >> <property> >> >> <name>dfs.dataname.data.dir</name> >> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/data</value> >> >> <final>true</final> >> >> </property> >> >> >> >> The java process are running on master as: >> >> 10479 SecondaryNameNode >> >> 10281 NameNode >> >> 10628 ResourceManager >> >> >> >> and on slave: >> >> 22870 DataNode >> >> 22991 NodeManager >> >> >> >> Any hints? Thanks! >> >> >> >> Cheers, >> >> Dan >> >> >> >> >> >> >> > >> > -- Regards, Jiayu (James) Ji, Cell: (312)823-7393