Thanks Jitendra, Bejoy and Yexi, I got past that. And now the ls command says it can not access the directory. I am sure this is a permissions issue. I am just wondering which directory and I missing permissions on.
Any pointers? And once again, thanks a lot Regards ashish *hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ hadoop fs -ls* *Warning: $HADOOP_HOME is deprecated.* * * *ls: Cannot access .: No such file or directory.* On Tue, Jul 23, 2013 at 9:42 AM, Jitendra Yadav <jeetuyadav200...@gmail.com>wrote: > Hi Ashish, > > Please check <property></property> in hdfs-site.xml. > > It is missing. > > Thanks. > On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani <ashish.umr...@gmail.com>wrote: > >> Hey thanks for response. I have changed 4 files during installation >> >> core-site.xml >> mapred-site.xml >> hdfs-site.xml and >> hadoop-env.sh >> >> >> I could not find any issues except that all params in the hadoop-env.sh >> are commented out. Only java_home is un commented. >> >> If you have a quick minute can you please browse through these files in >> email and let me know where could be the issue. >> >> Regards >> ashish >> >> >> >> I am listing those files below. >> *core-site.xml * >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <property> >> <name>hadoop.tmp.dir</name> >> <value>/app/hadoop/tmp</value> >> <description>A base for other temporary directories.</description> >> </property> >> >> <property> >> <name>fs.default.name</name> >> <value>hdfs://localhost:54310</value> >> <description>The name of the default file system. A URI whose >> scheme and authority determine the FileSystem implementation. The >> uri's scheme determines the config property (fs.SCHEME.impl) naming >> the FileSystem implementation class. The uri's authority is used to >> determine the host, port, etc. for a filesystem.</description> >> </property> >> </configuration> >> >> >> >> *mapred-site.xml* >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <property> >> <name>mapred.job.tracker</name> >> <value>localhost:54311</value> >> <description>The host and port that the MapReduce job tracker runs >> at. If "local", then jobs are run in-process as a single map >> and reduce task. >> </description> >> </property> >> </configuration> >> >> >> >> *hdfs-site.xml and* >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <name>dfs.replication</name> >> <value>1</value> >> <description>Default block replication. >> The actual number of replications can be specified when the file is >> created. >> The default is used if replication is not specified in create time. >> </description> >> </configuration> >> >> >> >> *hadoop-env.sh* >> # Set Hadoop-specific environment variables here. >> >> # The only required environment variable is JAVA_HOME. All others are >> # optional. When running a distributed configuration it is best to >> # set JAVA_HOME in this file, so that it is correctly defined on >> # remote nodes. >> >> # The java implementation to use. Required. >> export JAVA_HOME=/usr/lib/jvm/jdk1.7.0_25 >> >> # Extra Java CLASSPATH elements. Optional. >> # export HADOOP_CLASSPATH= >> >> >> All pther params in hadoop-env.sh are commented >> >> >> >> >> >> >> >> >> On Tue, Jul 23, 2013 at 8:38 AM, Jitendra Yadav < >> jeetuyadav200...@gmail.com> wrote: >> >>> Hi, >>> >>> You might have missed some configuration (XML tags ), Please check all >>> the Conf files. >>> >>> Thanks >>> On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani >>> <ashish.umr...@gmail.com>wrote: >>> >>>> Hi There, >>>> >>>> First of all, sorry if I am asking some stupid question. Myself being >>>> new to the Hadoop environment , am finding it a bit difficult to figure out >>>> why its failing >>>> >>>> I have installed hadoop 1.2, based on instructions given in the >>>> folllowing link >>>> >>>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >>>> >>>> All went well and I could do the start-all.sh and the jps command does >>>> show all 5 process to be present. >>>> >>>> However when I try to do >>>> >>>> hadoop fs -ls >>>> >>>> I get the following error >>>> >>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>> hadoop fs -ls >>>> Warning: $HADOOP_HOME is deprecated. >>>> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> ls: Cannot access .: No such file or directory. >>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>> >>>> >>>> >>>> Can someone help me figure out whats the issue in my installation >>>> >>>> >>>> Regards >>>> ashish >>>> >>> >>> >> >