Give complete hostname with domain name not just master-node. <property> <name>fs.defaultFS</name> <value>hdfs://master-node.domain.name:9000</value> </property>
Else give IP address also On 12/16/14, Dan Dong <dongda...@gmail.com> wrote: > Hi, Johny, > Yes, they have been turned off from the beginning. Guess the problem is > still in the conf files, it would be helpful if some example *.xml could be > shown. > > Cheers, > Dan > > > 2014-12-15 12:24 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >> >> do you have selinux and iptables turned off? >> >> ------------------------------ >> Date: Mon, 15 Dec 2014 09:54:41 -0600 >> Subject: Re: Hadoop 2.6.0: "FileSystem file:/// is not a distributed file >> system" >> From: dongda...@gmail.com >> To: user@hadoop.apache.org >> >> >> Found in the log file: >> 2014-12-12 15:51:10,434 ERROR >> org.apache.hadoop.hdfs.server.namenode.NameNode: >> java.lang.IllegalArgumentException: Does not contain a valid host:port >> authority: file:/// >> at >> org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:212) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:280) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:569) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1479) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1488) >> >> But I have set it in core-site.xml already: >> <property> >> <name>fs.defaultFS</name> >> <value>hdfs://master-node:9000</value> >> </property> >> >> Other settings: >> $ cat mapred-site.xml >> <configuration> >> <property> >> <name>mapred.job.tracker</name> >> <value>master-node:9002</value> >> </property> >> <property> >> <name>mapreduce.jobhistory.address</name> >> <value>master-node:10020</value> >> </property> >> <property> >> <name>mapreduce.jobhistory.webapp.address</name> >> <value>master-node:19888</value> >> </property> >> </configuration> >> >> $ cat yarn-site.xml >> <configuration> >> >> <!-- Site specific YARN configuration properties --> >> <property> >> <name>mapreduce.framework.name</name> >> <value>yarn</value> >> </property> >> <property> >> <name>yarn.resourcemanager.address</name> >> <value>master-node:18040</value> >> </property> >> <property> >> <name>yarn.resourcemanager.scheduler.address</name> >> <value>master-node:18030</value> >> </property> >> <property> >> <name>yarn.resourcemanager.webapp.address</name> >> <value>master-node:18088</value> >> </property> >> <property> >> <name>yarn.resourcemanager.resource-tracker.address</name> >> <value>master-node:18025</value> >> </property> >> <property> >> <name>yarn.resourcemanager.admin.address</name> >> <value>master-node:18141</value> >> </property> >> <property> >> <name>yarn.nodemanager.aux-services</name> >> <value>mapreduce_shuffle</value> >> </property> >> <property> >> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> >> <value>org.apache.hadoop.mapred.ShuffleHandler</value> >> </property> >> </configuration> >> >> Cheers, >> Dan >> >> >> 2014-12-15 9:17 GMT-06:00 Dan Dong <dongda...@gmail.com>: >> >> Thank you all, but still the same after change file:/ to file://, and >> HADOOP_CONF_DIR points to the correct position already: >> $ echo $HADOOP_CONF_DIR >> /home/dong/import/hadoop-2.6.0/etc/hadoop >> >> >> 2014-12-15 8:57 GMT-06:00 johny casanova <pcgamer2...@outlook.com>: >> >> Don't you have to use file:// instead of just one /? >> >> ------------------------------ >> From: brahmareddy.batt...@huawei.com >> To: user@hadoop.apache.org >> Subject: RE: Hadoop 2.6.0: "FileSystem file:/// is not a distributed file >> system" >> Date: Sat, 13 Dec 2014 05:48:18 +0000 >> >> >> Hi Dong, >> >> HADOOP_CONF_DIR might be referring to default..you can export >> HADOOP_CONF_DIR where following configuration files are present.. >> >> Thanks & Regards >> Brahma Reddy Battula >> >> >> ------------------------------ >> *From:* Dan Dong [dongda...@gmail.com] >> *Sent:* Saturday, December 13, 2014 3:43 AM >> *To:* user@hadoop.apache.org >> *Subject:* Hadoop 2.6.0: "FileSystem file:/// is not a distributed file >> system" >> >> Hi, >> I installed Hadoop2.6.0 on my cluster with 2 nodes, I got the following >> error when I run: >> $hadoop dfsadmin -report >> FileSystem file:/// is not a distributed file system >> >> What this mean? I have set it in core-site.xml already: >> <property> >> <name>fs.defaultFS</name> >> <value>hdfs://master-node:9000</value> >> </property> >> >> and in hdfs-site.xml: >> <property> >> <name>dfs.namenode.name.dir</name> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/name</value> >> <final>true</final> >> </property> >> <property> >> <name>dfs.dataname.data.dir</name> >> <value>file:/home/dong/hadoop-2.6.0-dist/dfs/data</value> >> <final>true</final> >> </property> >> >> The java process are running on master as: >> 10479 SecondaryNameNode >> 10281 NameNode >> 10628 ResourceManager >> >> and on slave: >> 22870 DataNode >> 22991 NodeManager >> >> Any hints? Thanks! >> >> Cheers, >> Dan >> >> >> >