Thanks Jitendra, Bejoy and Yexi,

I got past that.  And now the ls command says it can not access the
directory.  I am sure this is a permissions issue.  I am just wondering
which directory and I missing permissions on.

Any pointers?

And once again, thanks a lot

Regards
ashish

*hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$ hadoop
fs -ls*
*Warning: $HADOOP_HOME is deprecated.*
*
*
*ls: Cannot access .: No such file or directory.*



On Tue, Jul 23, 2013 at 9:42 AM, Jitendra Yadav
<jeetuyadav200...@gmail.com>wrote:

> Hi Ashish,
>
> Please check <property></property>  in hdfs-site.xml.
>
> It is missing.
>
> Thanks.
> On Tue, Jul 23, 2013 at 9:58 PM, Ashish Umrani <ashish.umr...@gmail.com>wrote:
>
>> Hey thanks for response.  I have changed 4 files during installation
>>
>> core-site.xml
>> mapred-site.xml
>> hdfs-site.xml   and
>> hadoop-env.sh
>>
>>
>> I could not find any issues except that all params in the hadoop-env.sh
>> are commented out.  Only java_home is un commented.
>>
>> If you have a quick minute can you please browse through these files in
>> email and let me know where could be the issue.
>>
>> Regards
>> ashish
>>
>>
>>
>> I am listing those files below.
>>  *core-site.xml *
>>  <?xml version="1.0"?>
>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>
>> <!-- Put site-specific property overrides in this file. -->
>>
>> <configuration>
>>   <property>
>>     <name>hadoop.tmp.dir</name>
>>     <value>/app/hadoop/tmp</value>
>>     <description>A base for other temporary directories.</description>
>>   </property>
>>
>>   <property>
>>     <name>fs.default.name</name>
>>     <value>hdfs://localhost:54310</value>
>>     <description>The name of the default file system.  A URI whose
>>     scheme and authority determine the FileSystem implementation.  The
>>     uri's scheme determines the config property (fs.SCHEME.impl) naming
>>     the FileSystem implementation class.  The uri's authority is used to
>>     determine the host, port, etc. for a filesystem.</description>
>>   </property>
>> </configuration>
>>
>>
>>
>> *mapred-site.xml*
>>  <?xml version="1.0"?>
>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>
>> <!-- Put site-specific property overrides in this file. -->
>>
>> <configuration>
>>   <property>
>>     <name>mapred.job.tracker</name>
>>     <value>localhost:54311</value>
>>     <description>The host and port that the MapReduce job tracker runs
>>     at.  If "local", then jobs are run in-process as a single map
>>     and reduce task.
>>     </description>
>>   </property>
>> </configuration>
>>
>>
>>
>> *hdfs-site.xml   and*
>>  <?xml version="1.0"?>
>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>>
>> <!-- Put site-specific property overrides in this file. -->
>>
>> <configuration>
>>   <name>dfs.replication</name>
>>   <value>1</value>
>>   <description>Default block replication.
>>     The actual number of replications can be specified when the file is
>> created.
>>     The default is used if replication is not specified in create time.
>>   </description>
>> </configuration>
>>
>>
>>
>> *hadoop-env.sh*
>>  # Set Hadoop-specific environment variables here.
>>
>> # The only required environment variable is JAVA_HOME.  All others are
>> # optional.  When running a distributed configuration it is best to
>> # set JAVA_HOME in this file, so that it is correctly defined on
>> # remote nodes.
>>
>> # The java implementation to use.  Required.
>> export JAVA_HOME=/usr/lib/jvm/jdk1.7.0_25
>>
>> # Extra Java CLASSPATH elements.  Optional.
>> # export HADOOP_CLASSPATH=
>>
>>
>> All pther params in hadoop-env.sh are commented
>>
>>
>>
>>
>>
>>
>>
>>
>> On Tue, Jul 23, 2013 at 8:38 AM, Jitendra Yadav <
>> jeetuyadav200...@gmail.com> wrote:
>>
>>> Hi,
>>>
>>> You might have missed some configuration (XML tags ), Please check all
>>> the Conf files.
>>>
>>> Thanks
>>> On Tue, Jul 23, 2013 at 6:25 PM, Ashish Umrani 
>>> <ashish.umr...@gmail.com>wrote:
>>>
>>>> Hi There,
>>>>
>>>> First of all, sorry if I am asking some stupid question.  Myself being
>>>> new to the Hadoop environment , am finding it a bit difficult to figure out
>>>> why its failing
>>>>
>>>> I have installed hadoop 1.2, based on instructions given in the
>>>> folllowing link
>>>>
>>>> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/
>>>>
>>>> All went well and I could do the start-all.sh and the jps command does
>>>> show all 5 process to be present.
>>>>
>>>> However when I try to do
>>>>
>>>> hadoop fs -ls
>>>>
>>>> I get the following error
>>>>
>>>>  hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$
>>>> hadoop fs -ls
>>>> Warning: $HADOOP_HOME is deprecated.
>>>>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> 13/07/23 05:55:06 WARN conf.Configuration: bad conf file: element not
>>>> <property>
>>>> ls: Cannot access .: No such file or directory.
>>>> hduser@ashish-HP-Pavilion-dv6-Notebook-PC:/usr/local/hadoop/conf$
>>>>
>>>>
>>>>
>>>> Can someone help me figure out whats the issue in my installation
>>>>
>>>>
>>>> Regards
>>>> ashish
>>>>
>>>
>>>
>>
>

Reply via email to