Hey, Thanks Shekhar. That worked like a chimp. Appreciate help from you all. Now I will try to put files and try the word count or similar program.
Regards ashish On Tue, Jul 23, 2013 at 10:07 AM, Shekhar Sharma <shekhar2...@gmail.com>wrote: > Its warning not error... > > Create a directory and then do ls ( In your case /user/hduser is not > created untill and unless for the first time you create a directory or put > some file) > > hadoop fs -mkdir sample > > hadoop fs -ls > > I would suggest if you are getting pemission problem, > please check the following: > > (1) Have you run the command "hadoop namenode -format" with different user > and you are accessing the hdfs with different user? > > On Tue, Jul 23, 2013 at 10:10 PM, <bejoy.had...@gmail.com> wrote: > >> ** >> Hi Ashish >> >> In your hdfs-site.xml within <configuration> tag you need to have the >> <property> tag and inside a <property> tag you can have <name>,<value> and >> <description> tags. >> >> Regards >> Bejoy KS >> >> Sent from remote device, Please excuse typos >> ------------------------------ >> *From: * Ashish Umrani <ashish.umr...@gmail.com> >> *Date: *Tue, 23 Jul 2013 09:28:00 -0700 >> *To: *<user@hadoop.apache.org> >> *ReplyTo: * user@hadoop.apache.org >> *Subject: *Re: New hadoop 1.2 single node installation giving problems >> >> Hey thanks for response. I have changed 4 files during installation >> >> core-site.xml >> mapred-site.xml >> hdfs-site.xml and >> hadoop-env.sh >> >> >> I could not find any issues except that all params in the hadoop-env.sh >> are commented out. Only java_home is un commented. >> >> If you have a quick minute can you please browse through these files in >> email and let me know where could be the issue. >> >> Regards >> ashish >> >> >> >> I am listing those files below. >> *core-site.xml * >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <property> >> <name>hadoop.tmp.dir</name> >> <value>/app/hadoop/tmp</value> >> <description>A base for other temporary directories.</description> >> </property> >> >> <property> >> <name>fs.default.name</name> >> <value>hdfs://localhost:54310</value> >> <description>The name of the default file system. A URI whose >> scheme and authority determine the FileSystem implementation. The >> uri's scheme determines the config property (fs.SCHEME.impl) naming >> the FileSystem implementation class. The uri's authority is used to >> determine the host, port, etc. for a filesystem.</description> >> </property> >> </configuration> >> >> >> >> *mapred-site.xml* >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <property> >> <name>mapred.job.tracker</name> >> <value>localhost:54311</value> >> <description>The host and port that the MapReduce job tracker runs >> at. If "local", then jobs are run in-process as a single map >> and reduce task. >> </description> >> </property> >> </configuration> >> >> >> >> *hdfs-site.xml and* >> <?xml version="1.0"?> >> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >> >> <!-- Put site-specific property overrides in this file. --> >> >> <configuration> >> <name>dfs.replication</name> >> <value>1</value> >> <description>Default block replication. >> The actual number of replications can be specified when the file is >> created. >> The default is used if replication is not specified in create time. >> </description> >> </configuration> >> >> >> >> *hadoop-env.sh* >> # Set Hadoop-specific environment variables here. >> >> # The only required environment variable is JAVA_HOME. All others are >> # optional. When running a distributed configuration it is best to >> # set JAVA_HOME in this file, so that it is correctly defined on >> # remote nodes. >> >> # The java implementation to use. Required. >> export JAVA_HOME=/usr/lib/jvm/jdk1.7.0_25 >> >> # Extra Java CLASSPATH elements. Optional. >> # export HADOOP_CLASSPATH= >> >> >> All pther params in hadoop-env.sh are commented >> >> >> >> >> >> >> >> >> On Tue, Jul 23, 2013 at 8:38 AM, Jitendra Yadav < >> jeetuyadav200...@gmail.com> wrote: >> >>> Hi, >>> >>> You might have missed some configuration (XML tags ), Please check all >>> the Conf files. >>> >>> Thanks >>> On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani >>> <ashish.umr...@gmail.com>wrote: >>> >>>> Hi There, >>>> >>>> First of all, sorry if I am asking some stupid question. Myself being >>>> new to the Hadoop environment , am finding it a bit difficult to figure out >>>> why its failing >>>> >>>> I have installed hadoop 1.2, based on instructions given in the >>>> folllowing link >>>> >>>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >>>> >>>> All went well and I could do the start-all.sh and the jps command does >>>> show all 5 process to be present. >>>> >>>> However when I try to do >>>> >>>> hadoop fs -ls >>>> >>>> I get the following error >>>> >>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>> hadoop fs -ls >>>> Warning: $HADOOP_HOME is deprecated. >>>> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not >>>> <property> >>>> ls: Cannot access .: No such file or directory. >>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ >>>> >>>> >>>> >>>> Can someone help me figure out whats the issue in my installation >>>> >>>> >>>> Regards >>>> ashish >>>> >>> >>> >> >