Hi James:
I verified that I have the following permission set for the path:

ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
total 4
drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
Thanks.
Richard


On Wed, Dec 8, 2010 at 4:50 PM, james warren <ja...@rockyou.com> wrote:

> Hi Richard -
>
> First thing that comes to mind is a permissions issue.  Can you verify that
> your directories along the desired namenode path are writable by the
> appropriate user(s)?
>
> HTH,
> -James
>
> On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang <richardtec...@gmail.com
> >wrote:
>
> > Hi Guys:
> > I am just installation the hadoop 0.21.0 in a single node cluster.
> > I encounter the following error when I run bin/hadoop namenode -format
> >
> > 10/12/08 16:27:22 ERROR namenode.NameNode:
> > java.io.IOException: Cannot create directory
> > /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
> >        at
> > org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
> >        at
> > org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
> >        at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
> >        at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
> >        at
> > org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
> >
> >
> > Below is my core-site.xml
> >
> > <configuration>
> > <!-- In: conf/core-site.xml -->
> > <property>
> >  <name>hadoop.tmp.dir</name>
> >  <value>/your/path/to/hadoop/tmp/dir/hadoop-${user.name}</value>
> >  <description>A base for other temporary directories.</description>
> > </property>
> >
> > <property>
> >  <name>fs.default.name</name>
> >  <value>hdfs://localhost:54310</value>
> >  <description>The name of the default file system.  A URI whose
> >  scheme and authority determine the FileSystem implementation.  The
> >  uri's scheme determines the config property (fs.SCHEME.impl) naming
> >  the FileSystem implementation class.  The uri's authority is used to
> >  determine the host, port, etc. for a filesystem.</description>
> > </property>
> > </configuration>
> >
> >
> > Below is my hdfs-site.xml
> > *<?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >
> > <!-- Put site-specific property overrides in this file. -->
> >
> > <configuration>
> > <!-- In: conf/hdfs-site.xml -->
> > <property>
> >  <name>dfs.replication</name>
> >  <value>1</value>
> >  <description>Default block replication.
> >  The actual number of replications can be specified when the file is
> > created.
> >  The default is used if replication is not specified in create time.
> >  </description>
> > </property>
> >
> > </configuration>
> >
> >
> > below is my mapred-site.xml:
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> >
> > <!-- Put site-specific property overrides in this file. -->
> >
> > <configuration>
> >
> > <!-- In: conf/mapred-site.xml -->
> > <property>
> >  <name>mapred.job.tracker</name>
> >  <value>localhost:54311</value>
> >  <description>The host and port that the MapReduce job tracker runs
> >  at.  If "local", then jobs are run in-process as a single map
> >  and reduce task.
> >  </description>
> > </property>
> >
> > </configuration>
> >
> >
> > Thanks.
> > Richard
> > *
> >
>

Reply via email to