Unfortunately in 1.x the format command's prompt is case-sensitive
(Fixed in 2.x):

You had:
Re-format filesystem in /app/hadoop/name ? (Y or N) y
Format aborted in /app/hadoop/name

Answer with a capital Y instead and it won't abort.

On Fri, Apr 27, 2012 at 3:07 PM, Mathias Schnydrig <smath...@ee.ethz.ch> wrote:
> Hi
>
> I am setting up a hadoop cluster with 5 slaves and a master, after the
> single node installation it worked fine, but after I went to a multinode
> cluster the namenode prints this message even after formatting the hadoop
> namenode: see below
> I also added the config files.
>
> The folders I am using exist on all nodes and the hadoop folder is placed
> also on all nodes in the same folder and the config files are all the same.
>
> I suppose it is some stupid error I did, as I am quite new to hadoop.
>
> Regards
> Mathias
>
> hduser@POISN-server:/usr/local/hadoop$ bin/hadoop namenode -format
> Warning: $HADOOP_HOME is deprecated.
>
> 12/04/27 10:39:52 INFO namenode.NameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting NameNode
> STARTUP_MSG:   host = POISN-server/127.0.1.1
> STARTUP_MSG:   args = [-format]
> STARTUP_MSG:   version = 1.0.2
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
> ************************************************************/
> Re-format filesystem in /app/hadoop/name ? (Y or N) y
> Format aborted in /app/hadoop/name
> 12/04/27 10:39:55 INFO namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at POISN-server/127.0.1.1
> ************************************************************/
> hduser@POISN-server:/usr/local/hadoop$ bin/hadoop namenode
> Warning: $HADOOP_HOME is deprecated.
>
> 12/04/27 10:40:04 INFO namenode.NameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting NameNode
> STARTUP_MSG:   host = POISN-server/127.0.1.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 1.0.2
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0.2 -r
> 1304954; compiled by 'hortonfo' on Sat Mar 24 23:58:21 UTC 2012
> ************************************************************/
> 12/04/27 10:40:04 INFO impl.MetricsConfig: loaded properties from
> hadoop-metrics2.properties
> 12/04/27 10:40:04 INFO impl.MetricsSourceAdapter: MBean for source
> MetricsSystem,sub=Stats registered.
> 12/04/27 10:40:04 INFO impl.MetricsSystemImpl: Scheduled snapshot period at
> 10 second(s).
> 12/04/27 10:40:04 INFO impl.MetricsSystemImpl: NameNode metrics system
> started
> 12/04/27 10:40:05 INFO impl.MetricsSourceAdapter: MBean for source ugi
> registered.
> 12/04/27 10:40:05 WARN impl.MetricsSystemImpl: Source name ugi already
> exists!
> 12/04/27 10:40:05 INFO impl.MetricsSourceAdapter: MBean for source jvm
> registered.
> 12/04/27 10:40:05 INFO impl.MetricsSourceAdapter: MBean for source NameNode
> registered.
> 12/04/27 10:40:05 INFO util.GSet: VM type       = 64-bit
> 12/04/27 10:40:05 INFO util.GSet: 2% max memory = 17.77875 MB
> 12/04/27 10:40:05 INFO util.GSet: capacity      = 2^21 = 2097152 entries
> 12/04/27 10:40:05 INFO util.GSet: recommended=2097152, actual=2097152
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: fsOwner=hduser
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: supergroup=supergroup
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: isPermissionEnabled=true
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: isAccessTokenEnabled=false
> accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
> 12/04/27 10:40:05 INFO namenode.FSNamesystem: Registered
> FSNamesystemStateMBean and NameNodeMXBean
> 12/04/27 10:40:05 INFO namenode.NameNode: Caching file names occuring more
> than 10 times
> 12/04/27 10:40:05 ERROR namenode.FSNamesystem: FSNamesystem initialization
> failed.
> java.io.IOException: NameNode is not formatted.
>    at
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:325)
>    at
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
> ...
>
> 12/04/27 10:40:05 INFO namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at POISN-server/127.0.1.1
> ************************************************************/
>
>
> hduser@POISN-server:/usr/local/hadoop/conf$ cat *-site.xml
>
> core-site.xml  hdfs-site.xml  mapred-site.xml
>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
> <name>fs.default.name</name>
> <value>hdfs://ClusterMaster:9000</value>
> </property>
> </configuration>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
> <property>
> <name>dfs.replication</name>
> <value>1</value>
> </property>
>
> <property>
> <name>dfs.name.dir</name>
> <value>/app/hadoop/name</value>
> </property>
>
> <property>
> <name>dfs.data.dir</name>
> <value>/app/hadoop/data</value>
> </property>
>
> </configuration>
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
>
> <!-- Put site-specific property overrides in this file. -->
>
> <configuration>
>
> <property>
> <name>mapred.job.tracker</name>
> <value>ClusterMaster:9001</value>
> </property>
>
> <property>
> <name>mapred.system.dir</name>
> <value>/app/hadoop/system</value>
> </property>
>
> <property>
> <name>mapred.local.dir</name>
> <value>/app/hadoop/local</value>
> </property>
>
> </configuration>
>
> taskcontroller.cfg:
> mapred.local.dir=/app/hadoop/local
> hadoop.log.dir=
> mapred.tasktracker.tasks.sleeptime-before-sigkill=#sleep time before sig
> kill is to be sent to process group after sigterm is sent. Should be in
> seconds
> mapreduce.tasktracker.group=hadoop
>
>



-- 
Harsh J

Reply via email to