oh, sorry. I corrected that typo hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name/current -l total 0 hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name -l total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 22:17 current
Even I remove the tmp I manually created and set all the Hadoop package to be 777. Then I run the hadoop again and it is still the same. Richard. On Wed, Dec 8, 2010 at 7:55 PM, Konstantin Boudnik <c...@apache.org> wrote: > Yeah, I figured that match. What I was referring to is the ending of the > paths: > .../hadoop-hadoop/dfs/name/current > .../hadoop-hadoop/dfs/hadoop > They are different > -- > Take care, > Konstantin (Cos) Boudnik > > > > On Wed, Dec 8, 2010 at 15:55, Richard Zhang <richardtec...@gmail.com> > wrote: > > Hi: > > "/your/path/to/hadoop" represents the location where hadoop is > installed. > > BTW, I believe this is a file writing permission problem. Because I use > the > > same *-site.xml setting to install with root and it works. > > But when I use the dedicated user hadoop, it always introduces this > problem. > > > > But I do created manually the directory path and grant with 755. > > Weird.... > > Richard. > > > > On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik <c...@apache.org> > wrote: > > > >> it seems that you are looking at 2 different directories: > >> > >> first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current > >> second: ls -l > tmp/dir/hadoop-hadoop/dfs/hadoop > >> -- > >> Take care, > >> Konstantin (Cos) Boudnik > >> > >> > >> > >> On Wed, Dec 8, 2010 at 14:19, Richard Zhang <richardtec...@gmail.com> > >> wrote: > >> > would that be the reason that 54310 port is not open? > >> > I just used > >> > * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT > >> > to open the port. > >> > But it seems the same erorr exists. > >> > Richard > >> > * > >> > On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang < > richardtec...@gmail.com > >> >wrote: > >> > > >> >> Hi James: > >> >> I verified that I have the following permission set for the path: > >> >> > >> >> ls -l tmp/dir/hadoop-hadoop/dfs/hadoop > >> >> total 4 > >> >> drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current > >> >> Thanks. > >> >> Richard > >> >> > >> >> > >> >> > >> >> On Wed, Dec 8, 2010 at 4:50 PM, james warren <ja...@rockyou.com> > wrote: > >> >> > >> >>> Hi Richard - > >> >>> > >> >>> First thing that comes to mind is a permissions issue. Can you > verify > >> >>> that > >> >>> your directories along the desired namenode path are writable by the > >> >>> appropriate user(s)? > >> >>> > >> >>> HTH, > >> >>> -James > >> >>> > >> >>> On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang < > richardtec...@gmail.com > >> >>> >wrote: > >> >>> > >> >>> > Hi Guys: > >> >>> > I am just installation the hadoop 0.21.0 in a single node cluster. > >> >>> > I encounter the following error when I run bin/hadoop namenode > >> -format > >> >>> > > >> >>> > 10/12/08 16:27:22 ERROR namenode.NameNode: > >> >>> > java.io.IOException: Cannot create directory > >> >>> > /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current > >> >>> > at > >> >>> > > >> >>> > > >> >>> > >> > org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) > >> >>> > at > >> >>> > > >> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) > >> >>> > at > >> >>> > > >> org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) > >> >>> > at > >> >>> > > >> >>> > >> > org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) > >> >>> > at > >> >>> > > >> >>> > > >> >>> > >> > org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) > >> >>> > at > >> >>> > > >> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) > >> >>> > > >> >>> > > >> >>> > Below is my core-site.xml > >> >>> > > >> >>> > <configuration> > >> >>> > <!-- In: conf/core-site.xml --> > >> >>> > <property> > >> >>> > <name>hadoop.tmp.dir</name> > >> >>> > <value>/your/path/to/hadoop/tmp/dir/hadoop-${user.name}</value> > >> >>> > <description>A base for other temporary > directories.</description> > >> >>> > </property> > >> >>> > > >> >>> > <property> > >> >>> > <name>fs.default.name</name> > >> >>> > <value>hdfs://localhost:54310</value> > >> >>> > <description>The name of the default file system. A URI whose > >> >>> > scheme and authority determine the FileSystem implementation. > The > >> >>> > uri's scheme determines the config property (fs.SCHEME.impl) > naming > >> >>> > the FileSystem implementation class. The uri's authority is used > to > >> >>> > determine the host, port, etc. for a filesystem.</description> > >> >>> > </property> > >> >>> > </configuration> > >> >>> > > >> >>> > > >> >>> > Below is my hdfs-site.xml > >> >>> > *<?xml version="1.0"?> > >> >>> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > >> >>> > > >> >>> > <!-- Put site-specific property overrides in this file. --> > >> >>> > > >> >>> > <configuration> > >> >>> > <!-- In: conf/hdfs-site.xml --> > >> >>> > <property> > >> >>> > <name>dfs.replication</name> > >> >>> > <value>1</value> > >> >>> > <description>Default block replication. > >> >>> > The actual number of replications can be specified when the file > is > >> >>> > created. > >> >>> > The default is used if replication is not specified in create > time. > >> >>> > </description> > >> >>> > </property> > >> >>> > > >> >>> > </configuration> > >> >>> > > >> >>> > > >> >>> > below is my mapred-site.xml: > >> >>> > <?xml version="1.0"?> > >> >>> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> > >> >>> > > >> >>> > <!-- Put site-specific property overrides in this file. --> > >> >>> > > >> >>> > <configuration> > >> >>> > > >> >>> > <!-- In: conf/mapred-site.xml --> > >> >>> > <property> > >> >>> > <name>mapred.job.tracker</name> > >> >>> > <value>localhost:54311</value> > >> >>> > <description>The host and port that the MapReduce job tracker > runs > >> >>> > at. If "local", then jobs are run in-process as a single map > >> >>> > and reduce task. > >> >>> > </description> > >> >>> > </property> > >> >>> > > >> >>> > </configuration> > >> >>> > > >> >>> > > >> >>> > Thanks. > >> >>> > Richard > >> >>> > * > >> >>> > > >> >>> > >> >> > >> >> > >> > > >> > > >