urgent, error: java.io.IOException: Cannot create directory
Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.comwrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.comwrote: Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
it seems that you are looking at 2 different directories: first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current second: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com wrote: would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.comwrote: Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
Hi: /your/path/to/hadoop represents the location where hadoop is installed. BTW, I believe this is a file writing permission problem. Because I use the same *-site.xml setting to install with root and it works. But when I use the dedicated user hadoop, it always introduces this problem. But I do created manually the directory path and grant with 755. Weird Richard. On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org wrote: it seems that you are looking at 2 different directories: first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current second: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com wrote: would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.com wrote: Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
Yeah, I figured that match. What I was referring to is the ending of the paths: .../hadoop-hadoop/dfs/name/current .../hadoop-hadoop/dfs/hadoop They are different -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 15:55, Richard Zhang richardtec...@gmail.com wrote: Hi: /your/path/to/hadoop represents the location where hadoop is installed. BTW, I believe this is a file writing permission problem. Because I use the same *-site.xml setting to install with root and it works. But when I use the dedicated user hadoop, it always introduces this problem. But I do created manually the directory path and grant with 755. Weird Richard. On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org wrote: it seems that you are looking at 2 different directories: first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current second: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com wrote: would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.com wrote: Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name valuelocalhost:54311/value descriptionThe host and port that the MapReduce job tracker runs at. If local, then jobs are run in-process as a single map and reduce task. /description /property /configuration Thanks. Richard *
Re: urgent, error: java.io.IOException: Cannot create directory
oh, sorry. I corrected that typo hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name/current -l total 0 hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name -l total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 22:17 current Even I remove the tmp I manually created and set all the Hadoop package to be 777. Then I run the hadoop again and it is still the same. Richard. On Wed, Dec 8, 2010 at 7:55 PM, Konstantin Boudnik c...@apache.org wrote: Yeah, I figured that match. What I was referring to is the ending of the paths: .../hadoop-hadoop/dfs/name/current .../hadoop-hadoop/dfs/hadoop They are different -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 15:55, Richard Zhang richardtec...@gmail.com wrote: Hi: /your/path/to/hadoop represents the location where hadoop is installed. BTW, I believe this is a file writing permission problem. Because I use the same *-site.xml setting to install with root and it works. But when I use the dedicated user hadoop, it always introduces this problem. But I do created manually the directory path and grant with 755. Weird Richard. On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org wrote: it seems that you are looking at 2 different directories: first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current second: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop -- Take care, Konstantin (Cos) Boudnik On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com wrote: would that be the reason that 54310 port is not open? I just used * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT to open the port. But it seems the same erorr exists. Richard * On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.com wrote: Hi James: I verified that I have the following permission set for the path: ls -l tmp/dir/hadoop-hadoop/dfs/hadoop total 4 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current Thanks. Richard On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote: Hi Richard - First thing that comes to mind is a permissions issue. Can you verify that your directories along the desired namenode path are writable by the appropriate user(s)? HTH, -James On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com wrote: Hi Guys: I am just installation the hadoop 0.21.0 in a single node cluster. I encounter the following error when I run bin/hadoop namenode -format 10/12/08 16:27:22 ERROR namenode.NameNode: java.io.IOException: Cannot create directory /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current at org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425) at org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444) at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368) Below is my core-site.xml configuration !-- In: conf/core-site.xml -- property namehadoop.tmp.dir/name value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value descriptionA base for other temporary directories./description /property property namefs.default.name/name valuehdfs://localhost:54310/value descriptionThe name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem./description /property /configuration Below is my hdfs-site.xml *?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/hdfs-site.xml -- property namedfs.replication/name value1/value descriptionDefault block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time. /description /property /configuration below is my mapred-site.xml: ?xml version=1.0? ?xml-stylesheet type=text/xsl href=configuration.xsl? !-- Put site-specific property overrides in this file. -- configuration !-- In: conf/mapred-site.xml -- property namemapred.job.tracker/name