urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Richard Zhang
Hi Guys:
I am just installation the hadoop 0.21.0 in a single node cluster.
I encounter the following error when I run bin/hadoop namenode -format

10/12/08 16:27:22 ERROR namenode.NameNode:
java.io.IOException: Cannot create directory
/your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
at
org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)


Below is my core-site.xml

configuration
!-- In: conf/core-site.xml --
property
  namehadoop.tmp.dir/name
  value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
  descriptionA base for other temporary directories./description
/property

property
  namefs.default.name/name
  valuehdfs://localhost:54310/value
  descriptionThe name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem./description
/property
/configuration


Below is my hdfs-site.xml
*?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?

!-- Put site-specific property overrides in this file. --

configuration
!-- In: conf/hdfs-site.xml --
property
  namedfs.replication/name
  value1/value
  descriptionDefault block replication.
  The actual number of replications can be specified when the file is
created.
  The default is used if replication is not specified in create time.
  /description
/property

/configuration


below is my mapred-site.xml:
?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?

!-- Put site-specific property overrides in this file. --

configuration

!-- In: conf/mapred-site.xml --
property
  namemapred.job.tracker/name
  valuelocalhost:54311/value
  descriptionThe host and port that the MapReduce job tracker runs
  at.  If local, then jobs are run in-process as a single map
  and reduce task.
  /description
/property

/configuration


Thanks.
Richard
*


Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread james warren
Hi Richard -

First thing that comes to mind is a permissions issue.  Can you verify that
your directories along the desired namenode path are writable by the
appropriate user(s)?

HTH,
-James

On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.comwrote:

 Hi Guys:
 I am just installation the hadoop 0.21.0 in a single node cluster.
 I encounter the following error when I run bin/hadoop namenode -format

 10/12/08 16:27:22 ERROR namenode.NameNode:
 java.io.IOException: Cannot create directory
 /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
at

 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
at
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
at
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
at
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
at

 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
at
 org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)


 Below is my core-site.xml

 configuration
 !-- In: conf/core-site.xml --
 property
  namehadoop.tmp.dir/name
  value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
  descriptionA base for other temporary directories./description
 /property

 property
  namefs.default.name/name
  valuehdfs://localhost:54310/value
  descriptionThe name of the default file system.  A URI whose
  scheme and authority determine the FileSystem implementation.  The
  uri's scheme determines the config property (fs.SCHEME.impl) naming
  the FileSystem implementation class.  The uri's authority is used to
  determine the host, port, etc. for a filesystem./description
 /property
 /configuration


 Below is my hdfs-site.xml
 *?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?

 !-- Put site-specific property overrides in this file. --

 configuration
 !-- In: conf/hdfs-site.xml --
 property
  namedfs.replication/name
  value1/value
  descriptionDefault block replication.
  The actual number of replications can be specified when the file is
 created.
  The default is used if replication is not specified in create time.
  /description
 /property

 /configuration


 below is my mapred-site.xml:
 ?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?

 !-- Put site-specific property overrides in this file. --

 configuration

 !-- In: conf/mapred-site.xml --
 property
  namemapred.job.tracker/name
  valuelocalhost:54311/value
  descriptionThe host and port that the MapReduce job tracker runs
  at.  If local, then jobs are run in-process as a single map
  and reduce task.
  /description
 /property

 /configuration


 Thanks.
 Richard
 *



Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Richard Zhang
Hi James:
I verified that I have the following permission set for the path:

ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
total 4
drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
Thanks.
Richard


On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote:

 Hi Richard -

 First thing that comes to mind is a permissions issue.  Can you verify that
 your directories along the desired namenode path are writable by the
 appropriate user(s)?

 HTH,
 -James

 On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com
 wrote:

  Hi Guys:
  I am just installation the hadoop 0.21.0 in a single node cluster.
  I encounter the following error when I run bin/hadoop namenode -format
 
  10/12/08 16:27:22 ERROR namenode.NameNode:
  java.io.IOException: Cannot create directory
  /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
 at
 
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
 at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
 at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
 at
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
 at
 
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
 at
  org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
 
 
  Below is my core-site.xml
 
  configuration
  !-- In: conf/core-site.xml --
  property
   namehadoop.tmp.dir/name
   value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
   descriptionA base for other temporary directories./description
  /property
 
  property
   namefs.default.name/name
   valuehdfs://localhost:54310/value
   descriptionThe name of the default file system.  A URI whose
   scheme and authority determine the FileSystem implementation.  The
   uri's scheme determines the config property (fs.SCHEME.impl) naming
   the FileSystem implementation class.  The uri's authority is used to
   determine the host, port, etc. for a filesystem./description
  /property
  /configuration
 
 
  Below is my hdfs-site.xml
  *?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
  !-- In: conf/hdfs-site.xml --
  property
   namedfs.replication/name
   value1/value
   descriptionDefault block replication.
   The actual number of replications can be specified when the file is
  created.
   The default is used if replication is not specified in create time.
   /description
  /property
 
  /configuration
 
 
  below is my mapred-site.xml:
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
 
  !-- In: conf/mapred-site.xml --
  property
   namemapred.job.tracker/name
   valuelocalhost:54311/value
   descriptionThe host and port that the MapReduce job tracker runs
   at.  If local, then jobs are run in-process as a single map
   and reduce task.
   /description
  /property
 
  /configuration
 
 
  Thanks.
  Richard
  *
 



Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Richard Zhang
would that be the reason that 54310 port is not open?
I just used
* iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
to open the port.
But it seems the same erorr exists.
Richard
*
On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.comwrote:

 Hi James:
 I verified that I have the following permission set for the path:

 ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
 total 4
 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
 Thanks.
 Richard



 On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote:

 Hi Richard -

 First thing that comes to mind is a permissions issue.  Can you verify
 that
 your directories along the desired namenode path are writable by the
 appropriate user(s)?

 HTH,
 -James

 On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com
 wrote:

  Hi Guys:
  I am just installation the hadoop 0.21.0 in a single node cluster.
  I encounter the following error when I run bin/hadoop namenode -format
 
  10/12/08 16:27:22 ERROR namenode.NameNode:
  java.io.IOException: Cannot create directory
  /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
 at
 
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
 at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
 at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
 at
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
 at
 
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
 at
  org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
 
 
  Below is my core-site.xml
 
  configuration
  !-- In: conf/core-site.xml --
  property
   namehadoop.tmp.dir/name
   value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
   descriptionA base for other temporary directories./description
  /property
 
  property
   namefs.default.name/name
   valuehdfs://localhost:54310/value
   descriptionThe name of the default file system.  A URI whose
   scheme and authority determine the FileSystem implementation.  The
   uri's scheme determines the config property (fs.SCHEME.impl) naming
   the FileSystem implementation class.  The uri's authority is used to
   determine the host, port, etc. for a filesystem./description
  /property
  /configuration
 
 
  Below is my hdfs-site.xml
  *?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
  !-- In: conf/hdfs-site.xml --
  property
   namedfs.replication/name
   value1/value
   descriptionDefault block replication.
   The actual number of replications can be specified when the file is
  created.
   The default is used if replication is not specified in create time.
   /description
  /property
 
  /configuration
 
 
  below is my mapred-site.xml:
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
 
  !-- In: conf/mapred-site.xml --
  property
   namemapred.job.tracker/name
   valuelocalhost:54311/value
   descriptionThe host and port that the MapReduce job tracker runs
   at.  If local, then jobs are run in-process as a single map
   and reduce task.
   /description
  /property
 
  /configuration
 
 
  Thanks.
  Richard
  *
 





Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Konstantin Boudnik
it seems that you are looking at 2 different directories:

first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
second: ls -l  tmp/dir/hadoop-hadoop/dfs/hadoop
--
  Take care,
Konstantin (Cos) Boudnik



On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com wrote:
 would that be the reason that 54310 port is not open?
 I just used
 * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
 to open the port.
 But it seems the same erorr exists.
 Richard
 *
 On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.comwrote:

 Hi James:
 I verified that I have the following permission set for the path:

 ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
 total 4
 drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
 Thanks.
 Richard



 On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote:

 Hi Richard -

 First thing that comes to mind is a permissions issue.  Can you verify
 that
 your directories along the desired namenode path are writable by the
 appropriate user(s)?

 HTH,
 -James

 On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com
 wrote:

  Hi Guys:
  I am just installation the hadoop 0.21.0 in a single node cluster.
  I encounter the following error when I run bin/hadoop namenode -format
 
  10/12/08 16:27:22 ERROR namenode.NameNode:
  java.io.IOException: Cannot create directory
  /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
         at
 
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
         at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
         at
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
         at
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
         at
 
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
         at
  org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
 
 
  Below is my core-site.xml
 
  configuration
  !-- In: conf/core-site.xml --
  property
   namehadoop.tmp.dir/name
   value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
   descriptionA base for other temporary directories./description
  /property
 
  property
   namefs.default.name/name
   valuehdfs://localhost:54310/value
   descriptionThe name of the default file system.  A URI whose
   scheme and authority determine the FileSystem implementation.  The
   uri's scheme determines the config property (fs.SCHEME.impl) naming
   the FileSystem implementation class.  The uri's authority is used to
   determine the host, port, etc. for a filesystem./description
  /property
  /configuration
 
 
  Below is my hdfs-site.xml
  *?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
  !-- In: conf/hdfs-site.xml --
  property
   namedfs.replication/name
   value1/value
   descriptionDefault block replication.
   The actual number of replications can be specified when the file is
  created.
   The default is used if replication is not specified in create time.
   /description
  /property
 
  /configuration
 
 
  below is my mapred-site.xml:
  ?xml version=1.0?
  ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
  !-- Put site-specific property overrides in this file. --
 
  configuration
 
  !-- In: conf/mapred-site.xml --
  property
   namemapred.job.tracker/name
   valuelocalhost:54311/value
   descriptionThe host and port that the MapReduce job tracker runs
   at.  If local, then jobs are run in-process as a single map
   and reduce task.
   /description
  /property
 
  /configuration
 
 
  Thanks.
  Richard
  *
 






Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Richard Zhang
Hi:
/your/path/to/hadoop  represents the location where hadoop is installed.
BTW, I believe this is a file writing permission problem. Because I use the
same *-site.xml setting to install with root and it works.
But when I use the dedicated user hadoop, it always introduces this problem.

But I do created manually the directory path and grant with 755.
Weird
Richard.

On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org wrote:

 it seems that you are looking at 2 different directories:

 first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
 second: ls -l  tmp/dir/hadoop-hadoop/dfs/hadoop
 --
   Take care,
 Konstantin (Cos) Boudnik



 On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com
 wrote:
  would that be the reason that 54310 port is not open?
  I just used
  * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
  to open the port.
  But it seems the same erorr exists.
  Richard
  *
  On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.com
 wrote:
 
  Hi James:
  I verified that I have the following permission set for the path:
 
  ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
  total 4
  drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
  Thanks.
  Richard
 
 
 
  On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote:
 
  Hi Richard -
 
  First thing that comes to mind is a permissions issue.  Can you verify
  that
  your directories along the desired namenode path are writable by the
  appropriate user(s)?
 
  HTH,
  -James
 
  On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com
  wrote:
 
   Hi Guys:
   I am just installation the hadoop 0.21.0 in a single node cluster.
   I encounter the following error when I run bin/hadoop namenode
 -format
  
   10/12/08 16:27:22 ERROR namenode.NameNode:
   java.io.IOException: Cannot create directory
   /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
  at
  
  
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
  at
  
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
  at
  
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
  at
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
  at
  
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
  at
  
 org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
  
  
   Below is my core-site.xml
  
   configuration
   !-- In: conf/core-site.xml --
   property
namehadoop.tmp.dir/name
value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
descriptionA base for other temporary directories./description
   /property
  
   property
namefs.default.name/name
valuehdfs://localhost:54310/value
descriptionThe name of the default file system.  A URI whose
scheme and authority determine the FileSystem implementation.  The
uri's scheme determines the config property (fs.SCHEME.impl) naming
the FileSystem implementation class.  The uri's authority is used to
determine the host, port, etc. for a filesystem./description
   /property
   /configuration
  
  
   Below is my hdfs-site.xml
   *?xml version=1.0?
   ?xml-stylesheet type=text/xsl href=configuration.xsl?
  
   !-- Put site-specific property overrides in this file. --
  
   configuration
   !-- In: conf/hdfs-site.xml --
   property
namedfs.replication/name
value1/value
descriptionDefault block replication.
The actual number of replications can be specified when the file is
   created.
The default is used if replication is not specified in create time.
/description
   /property
  
   /configuration
  
  
   below is my mapred-site.xml:
   ?xml version=1.0?
   ?xml-stylesheet type=text/xsl href=configuration.xsl?
  
   !-- Put site-specific property overrides in this file. --
  
   configuration
  
   !-- In: conf/mapred-site.xml --
   property
namemapred.job.tracker/name
valuelocalhost:54311/value
descriptionThe host and port that the MapReduce job tracker runs
at.  If local, then jobs are run in-process as a single map
and reduce task.
/description
   /property
  
   /configuration
  
  
   Thanks.
   Richard
   *
  
 
 
 
 



Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Konstantin Boudnik
Yeah, I figured that match. What I was referring to is the ending of the paths:
.../hadoop-hadoop/dfs/name/current
.../hadoop-hadoop/dfs/hadoop
They are different
--
  Take care,
Konstantin (Cos) Boudnik



On Wed, Dec 8, 2010 at 15:55, Richard Zhang richardtec...@gmail.com wrote:
 Hi:
 /your/path/to/hadoop  represents the location where hadoop is installed.
 BTW, I believe this is a file writing permission problem. Because I use the
 same *-site.xml setting to install with root and it works.
 But when I use the dedicated user hadoop, it always introduces this problem.

 But I do created manually the directory path and grant with 755.
 Weird
 Richard.

 On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org wrote:

 it seems that you are looking at 2 different directories:

 first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
 second: ls -l                              tmp/dir/hadoop-hadoop/dfs/hadoop
 --
   Take care,
 Konstantin (Cos) Boudnik



 On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com
 wrote:
  would that be the reason that 54310 port is not open?
  I just used
  * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
  to open the port.
  But it seems the same erorr exists.
  Richard
  *
  On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang richardtec...@gmail.com
 wrote:
 
  Hi James:
  I verified that I have the following permission set for the path:
 
  ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
  total 4
  drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
  Thanks.
  Richard
 
 
 
  On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com wrote:
 
  Hi Richard -
 
  First thing that comes to mind is a permissions issue.  Can you verify
  that
  your directories along the desired namenode path are writable by the
  appropriate user(s)?
 
  HTH,
  -James
 
  On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang richardtec...@gmail.com
  wrote:
 
   Hi Guys:
   I am just installation the hadoop 0.21.0 in a single node cluster.
   I encounter the following error when I run bin/hadoop namenode
 -format
  
   10/12/08 16:27:22 ERROR namenode.NameNode:
   java.io.IOException: Cannot create directory
   /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
          at
  
  
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
          at
  
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
          at
  
 org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
          at
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
          at
  
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
          at
  
 org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
  
  
   Below is my core-site.xml
  
   configuration
   !-- In: conf/core-site.xml --
   property
    namehadoop.tmp.dir/name
    value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
    descriptionA base for other temporary directories./description
   /property
  
   property
    namefs.default.name/name
    valuehdfs://localhost:54310/value
    descriptionThe name of the default file system.  A URI whose
    scheme and authority determine the FileSystem implementation.  The
    uri's scheme determines the config property (fs.SCHEME.impl) naming
    the FileSystem implementation class.  The uri's authority is used to
    determine the host, port, etc. for a filesystem./description
   /property
   /configuration
  
  
   Below is my hdfs-site.xml
   *?xml version=1.0?
   ?xml-stylesheet type=text/xsl href=configuration.xsl?
  
   !-- Put site-specific property overrides in this file. --
  
   configuration
   !-- In: conf/hdfs-site.xml --
   property
    namedfs.replication/name
    value1/value
    descriptionDefault block replication.
    The actual number of replications can be specified when the file is
   created.
    The default is used if replication is not specified in create time.
    /description
   /property
  
   /configuration
  
  
   below is my mapred-site.xml:
   ?xml version=1.0?
   ?xml-stylesheet type=text/xsl href=configuration.xsl?
  
   !-- Put site-specific property overrides in this file. --
  
   configuration
  
   !-- In: conf/mapred-site.xml --
   property
    namemapred.job.tracker/name
    valuelocalhost:54311/value
    descriptionThe host and port that the MapReduce job tracker runs
    at.  If local, then jobs are run in-process as a single map
    and reduce task.
    /description
   /property
  
   /configuration
  
  
   Thanks.
   Richard
   *
  
 
 
 
 




Re: urgent, error: java.io.IOException: Cannot create directory

2010-12-08 Thread Richard Zhang
oh, sorry. I corrected that typo
hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name/current -l
total 0
hadoop$ ls tmp/dir/hadoop-hadoop/dfs/name -l
total 4
drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 22:17 current

Even I remove the tmp I manually created and set all the Hadoop package to
be 777. Then I run the hadoop again and it is still the same.

Richard.

On Wed, Dec 8, 2010 at 7:55 PM, Konstantin Boudnik c...@apache.org wrote:

 Yeah, I figured that match. What I was referring to is the ending of the
 paths:
 .../hadoop-hadoop/dfs/name/current
 .../hadoop-hadoop/dfs/hadoop
 They are different
 --
   Take care,
 Konstantin (Cos) Boudnik



 On Wed, Dec 8, 2010 at 15:55, Richard Zhang richardtec...@gmail.com
 wrote:
  Hi:
  /your/path/to/hadoop  represents the location where hadoop is
 installed.
  BTW, I believe this is a file writing permission problem. Because I use
 the
  same *-site.xml setting to install with root and it works.
  But when I use the dedicated user hadoop, it always introduces this
 problem.
 
  But I do created manually the directory path and grant with 755.
  Weird
  Richard.
 
  On Wed, Dec 8, 2010 at 6:51 PM, Konstantin Boudnik c...@apache.org
 wrote:
 
  it seems that you are looking at 2 different directories:
 
  first post: /your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
  second: ls -l
  tmp/dir/hadoop-hadoop/dfs/hadoop
  --
Take care,
  Konstantin (Cos) Boudnik
 
 
 
  On Wed, Dec 8, 2010 at 14:19, Richard Zhang richardtec...@gmail.com
  wrote:
   would that be the reason that 54310 port is not open?
   I just used
   * iptables -A INPUT -p tcp --dport 54310 -j ACCEPT
   to open the port.
   But it seems the same erorr exists.
   Richard
   *
   On Wed, Dec 8, 2010 at 4:56 PM, Richard Zhang 
 richardtec...@gmail.com
  wrote:
  
   Hi James:
   I verified that I have the following permission set for the path:
  
   ls -l tmp/dir/hadoop-hadoop/dfs/hadoop
   total 4
   drwxr-xr-x 2 hadoop hadoop 4096 2010-12-08 15:56 current
   Thanks.
   Richard
  
  
  
   On Wed, Dec 8, 2010 at 4:50 PM, james warren ja...@rockyou.com
 wrote:
  
   Hi Richard -
  
   First thing that comes to mind is a permissions issue.  Can you
 verify
   that
   your directories along the desired namenode path are writable by the
   appropriate user(s)?
  
   HTH,
   -James
  
   On Wed, Dec 8, 2010 at 1:37 PM, Richard Zhang 
 richardtec...@gmail.com
   wrote:
  
Hi Guys:
I am just installation the hadoop 0.21.0 in a single node cluster.
I encounter the following error when I run bin/hadoop namenode
  -format
   
10/12/08 16:27:22 ERROR namenode.NameNode:
java.io.IOException: Cannot create directory
/your/path/to/hadoop/tmp/dir/hadoop-hadoop/dfs/name/current
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.common.Storage$StorageDirectory.clearDirectory(Storage.java:312)
   at
   
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1425)
   at
   
  org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:1444)
   at
   
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1242)
   at
   
   
  
 
 org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1348)
   at
   
  org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1368)
   
   
Below is my core-site.xml
   
configuration
!-- In: conf/core-site.xml --
property
 namehadoop.tmp.dir/name
 value/your/path/to/hadoop/tmp/dir/hadoop-${user.name}/value
 descriptionA base for other temporary
 directories./description
/property
   
property
 namefs.default.name/name
 valuehdfs://localhost:54310/value
 descriptionThe name of the default file system.  A URI whose
 scheme and authority determine the FileSystem implementation.
  The
 uri's scheme determines the config property (fs.SCHEME.impl)
 naming
 the FileSystem implementation class.  The uri's authority is used
 to
 determine the host, port, etc. for a filesystem./description
/property
/configuration
   
   
Below is my hdfs-site.xml
*?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?
   
!-- Put site-specific property overrides in this file. --
   
configuration
!-- In: conf/hdfs-site.xml --
property
 namedfs.replication/name
 value1/value
 descriptionDefault block replication.
 The actual number of replications can be specified when the file
 is
created.
 The default is used if replication is not specified in create
 time.
 /description
/property
   
/configuration
   
   
below is my mapred-site.xml:
?xml version=1.0?
?xml-stylesheet type=text/xsl href=configuration.xsl?
   
!-- Put site-specific property overrides in this file. --
   
configuration
   
!-- In: conf/mapred-site.xml --
property
 namemapred.job.tracker/name