Hi Ashish, Please check <property></property> in hdfs-site.xml.
It is missing. Thanks. On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani <ashish.umr...@gmail.com>wrote: > Hey thanks for response. I have changed 4 files during installation > > core-site.xml > mapred-site.xml > hdfs-site.xml and > hadoop-env.sh > > > I could not find any issues except that all params in the hadoop-env.sh > are commented out. Only java_home is un commented. > > If you have a quick minute can you please browse through these files in > email and let me know where could be the issue. > > Regards > ashish > > > > I am listing those files below. > *core-site.xml * > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <property> > <name>hadoop.tmp.dir</name> > <value>/app/hadoop/tmp</value> > <description>A base for other temporary directories.</description> > </property> > > <property> > <name>fs.default.name</name> > <value>hdfs://localhost:54310</value> > <description>The name of the default file system. A URI whose > scheme and authority determine the FileSystem implementation. The > uri's scheme determines the config property (fs.SCHEME.impl) naming > the FileSystem implementation class. The uri's authority is used to > determine the host, port, etc. for a filesystem.</description> > </property> > </configuration> > > > > *mapred-site.xml* > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <property> > <name>mapred.job.tracker</name> > <value>localhost:54311</value> > <description>The host and port that the MapReduce job tracker runs > at. If "local", then jobs are run in-process as a single map > and reduce task. > </description> > </property> > </configuration> > > > > *hdfs-site.xml and* > <?xml version="1.0"?> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > > <!-- Put site-specific property overrides in this file. --> > > <configuration> > <name>dfs.replication</name> > <value>1</value> > <description>Default block replication. > The actual number of replications can be specified when the file is > created. > The default is used if replication is not specified in create time. > </description> > </configuration> > > > > *hadoop-env.sh* > # Set Hadoop-specific environment variables here. > > # The only required environment variable is JAVA_HOME. All others are > # optional. When running a distributed configuration it is best to > # set JAVA_HOME in this file, so that it is correctly defined on > # remote nodes. > > # The java implementation to use. Required. > export JAVA_HOME=/usr/lib/jvm/jdk1.7.0_25 > > # Extra Java CLASSPATH elements. Optional. > # export HADOOP_CLASSPATH= > > > All pther params in hadoop-env.sh are commented > > > > > > > > > On Tue, Jul 23, 2013 at 8:38 AM, Jitendra Yadav < > jeetuyadav200...@gmail.com> wrote: > >> Hi, >> >> You might have missed some configuration (XML tags ), Please check all >> the Conf files. >> >> Thanks >> On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani >> <ashish.umr...@gmail.com>wrote: >> >>> Hi There, >>> >>> First of all, sorry if I am asking some stupid question. Myself being >>> new to the Hadoop environment , am finding it a bit difficult to figure out >>> why its failing >>> >>> I have installed hadoop 1.2, based on instructions given in the >>> folllowing link >>> >>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >>> >>> All went well and I could do the start-all.sh and the jps command does >>> show all 5 process to be present. >>> >>> However when I try to do >>> >>> hadoop fs -ls >>> >>> I get the following error >>> >>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>> hadoop fs -ls >>> Warning: $HADOOP_HOME is deprecated. >>> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>> <property> >>> ls: Cannot access .: No such file or directory. >>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>> >>> >>> >>> Can someone help me figure out whats the issue in my installation >>> >>> >>> Regards >>> ashish >>> >> >> >