[ https://issues.apache.org/jira/browse/HDFS-100?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Ravi Phulari resolved HDFS-100. ------------------------------- Resolution: Invalid This is more of hadoop format and cluster set up issue, not a bug. Please post problems related to startup or usage on user mailing list. > dfs startup error, 0 datanodes in > ---------------------------------- > > Key: HDFS-100 > URL: https://issues.apache.org/jira/browse/HDFS-100 > Project: Hadoop HDFS > Issue Type: Bug > Environment: hadoop-site.xml setting: > fs.default.name hdfs://master.cloud:9000 > mapred.job.tracker hdfs://master.cloud:9001 > hadoop.tmp.dir /home/user/hadoop/tmp/ > mapred.chile.java.opts Xmls512M > Reporter: Focus > > Web site shows: > NameNode 'master.cloud:9000' > Started: Thu Dec 18 17:10:35 CST 2008 > Version: 0.17.2.1, r684969 > Compiled: Wed Aug 20 22:29:32 UTC 2008 by oom > Upgrades: There are no upgrades in progress. > Browse the filesystem > -------------------------------------------------------------------------------- > Cluster Summary > Safe mode is ON. The ratio of reported blocks 0.0000 has not reached the > threshold 0.9990. Safe mode will be turned off automatically. > 21 files and directories, 6 blocks = 27 total. Heap Size is 4.94 MB / 992.31 > MB (0%) > Capacity : 0 KB > DFS Remaining : 0 KB > DFS Used : 0 KB > DFS Used% : 0 % > Live Nodes : 0 > Dead Nodes : 0 > -------------------------------------------------------------------------------- > There are no datanodes in the cluster > On blog of namenode, it shows: > 2008-12-18 17:10:35,204 INFO org.apache.hadoop.dfs.NameNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting NameNode > STARTUP_MSG: host = master.cloud/10.100.4.226 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.17.2.1 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.17 -r 684969; > compiled by 'oom' on Wed Aug 20 22:29:32 UTC 2008 > ************************************************************/ > 2008-12-18 17:10:35,337 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: > Initializing RPC Metrics with hostName=NameNode, port=9000 > 2008-12-18 17:10:35,344 INFO org.apache.hadoop.dfs.NameNode: Namenode up at: > master.cloud/10.100.4.226:9000 > 2008-12-18 17:10:35,348 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=NameNode, sessionId=null > 2008-12-18 17:10:35,351 INFO org.apache.hadoop.dfs.NameNodeMetrics: > Initializing NameNodeMeterics using context > object:org.apache.hadoop.metrics.spi.NullContext > 2008-12-18 17:10:35,436 INFO org.apache.hadoop.fs.FSNamesystem: > fsOwner=user,users,ftp,sshd > 2008-12-18 17:10:35,437 INFO org.apache.hadoop.fs.FSNamesystem: > supergroup=supergroup > 2008-12-18 17:10:35,437 INFO org.apache.hadoop.fs.FSNamesystem: > isPermissionEnabled=true > 2008-12-18 17:10:35,576 INFO org.apache.hadoop.fs.FSNamesystem: Finished > loading FSImage in 181 msecs > 2008-12-18 17:10:35,585 INFO org.apache.hadoop.dfs.StateChange: STATE* Safe > mode ON. > The ratio of reported blocks 0.0000 has not reached the threshold 0.9990. > Safe mode will be turned off automatically. > 2008-12-18 17:10:35,595 INFO org.apache.hadoop.fs.FSNamesystem: Registered > FSNamesystemStatusMBean > 2008-12-18 17:10:35,727 INFO org.mortbay.util.Credential: Checking Resource > aliases > 2008-12-18 17:10:35,870 INFO org.mortbay.http.HttpServer: Version Jetty/5.1.4 > 2008-12-18 17:10:35,871 INFO org.mortbay.util.Container: Started > HttpContext[/static,/static] > 2008-12-18 17:10:35,871 INFO org.mortbay.util.Container: Started > HttpContext[/logs,/logs] > 2008-12-18 17:10:36,260 INFO org.mortbay.util.Container: Started > org.mortbay.jetty.servlet.webapplicationhand...@b60b93 > 2008-12-18 17:10:36,307 INFO org.mortbay.util.Container: Started > WebApplicationContext[/,/] > 2008-12-18 17:10:36,309 INFO org.mortbay.http.SocketListener: Started > SocketListener on 0.0.0.0:50070 > 2008-12-18 17:10:36,310 INFO org.mortbay.util.Container: Started > org.mortbay.jetty.ser...@1bd7848 > 2008-12-18 17:10:36,310 INFO org.apache.hadoop.fs.FSNamesystem: Web-server up > at: 0.0.0.0:50070 > 2008-12-18 17:10:36,310 INFO org.apache.hadoop.ipc.Server: IPC Server > Responder: starting > 2008-12-18 17:10:36,312 INFO org.apache.hadoop.ipc.Server: IPC Server > listener on 9000: starting > 2008-12-18 17:10:36,316 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 0 on 9000: starting > 2008-12-18 17:10:36,317 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 2 on 9000: starting > 2008-12-18 17:10:36,317 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 3 on 9000: starting > 2008-12-18 17:10:36,320 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 4 on 9000: starting > 2008-12-18 17:10:36,321 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 5 on 9000: starting > 2008-12-18 17:10:36,321 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 6 on 9000: starting > 2008-12-18 17:10:36,321 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 7 on 9000: starting > 2008-12-18 17:10:36,321 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 1 on 9000: starting > 2008-12-18 17:10:36,322 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 8 on 9000: starting > 2008-12-18 17:10:36,374 INFO org.apache.hadoop.ipc.Server: IPC Server handler > 9 on 9000: starting > and in the slaves blog, i find a strange thing. > ************************************************************/ > 2008-12-18 17:11:47,627 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = slave3.cloud/127.0.0.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.17.2.1 > STARTUP_MSG: build = > https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.17 -r 684969; > compiled by 'oom' on Wed Aug 20 22:29:32 UTC 2008 > ************************************************************/ > 2008-12-18 17:11:48,267 ERROR org.apache.hadoop.dfs.DataNode: > java.io.IOException: Incompatible namespaceIDs in > /home/user/hadoop/tmp/dfs/data: namenode namespaceID = 1098832880; datanode > namespaceID = 464592288 > at > org.apache.hadoop.dfs.DataStorage.doTransition(DataStorage.java:298) > at > org.apache.hadoop.dfs.DataStorage.recoverTransitionRead(DataStorage.java:142) > at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:258) > at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:176) > at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2795) > at > org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2750) > at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2758) > at org.apache.hadoop.dfs.DataNode.main(DataNode.java:2880) > 2008-12-18 17:11:48,269 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG: > /************************************************************ > SHUTDOWN_MSG: Shutting down DataNode at slave3.cloud/127.0.0.1 > ************************************************************/ -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.