Hello Jon, Could you please verify that your node can resolve the host name?
It would be helpful too if you can attach your configuration files and the output of: HADOOP_ROOT_LOGGER=DEBUG,console hadoop fs -ls / as Todd suggested. Cheers, esteban On Jan 1, 2011 2:01 PM, "Jon Lederman" <jon2...@gmail.com> wrote: > Hi, > > Still no luck in getting FS commands to work. I did take a look at the logs. They all look pretty clean with the following exceptions: The DataNode appears to start up fine. However, the NameNode reports that the Network Topology has 0 racks and 0 datanodes. Is this normal? Is it possible the namenode cannot talk to the datanode? Any thoughts on what might be wrong? > > Thanks in advance and happy new year. > > -Jon > 2011-01-01 19:45:27,197 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = localhost/127.0.0.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.20.2 > STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled > by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 > ************************************************************/ > sc-ssh-svr1 logs $ more hadoop-root-namenode-localhost.log > 2011-01-01 19:45:23,988 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting NameNode > STARTUP_MSG: host = localhost/127.0.0.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.20.2 > STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled > by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 > ************************************************************/ > 2011-01-01 19:45:27,059 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics with hostName= > NameNode, port=8020 > 2011-01-01 19:45:28,355 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: localhost.locald > omain/127.0.0.1:8020 > 2011-01-01 19:45:28,448 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics with processNa > me=NameNode, sessionId=null > 2011-01-01 19:45:28,492 INFO org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing Name > NodeMeterics using context object:org.apache.hadoop.metrics.spi.NullContext > 2011-01-01 19:45:29,758 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner=root,root > 2011-01-01 19:45:29,763 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup > 2011-01-01 19:45:29,770 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: isPermissionEnabled=true > 2011-01-01 19:45:29,965 INFO org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: Initializing > FSNamesystemMetrics using context object:org.apache.hadoop.metrics.spi.NullContext > 2011-01-01 19:45:29,994 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered FSNamesystemStatu > sMBean > 2011-01-01 19:45:30,603 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files = 1 > 2011-01-01 19:45:30,696 INFO org.apache.hadoop.hdfs.server.common.Storage: Number of files under construction > = 0 > 2011-01-01 19:45:30,701 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 loaded in 0 s > econds. > 2011-01-01 19:45:30,708 INFO org.apache.hadoop.hdfs.server.common.Storage: Edits file /tmp/hadoop-root/dfs/nam > e/current/edits of size 4 edits # 0 loaded in 0 seconds. > 2011-01-01 19:45:30,767 INFO org.apache.hadoop.hdfs.server.common.Storage: Image file of size 94 saved in 0 se > conds. > 2011-01-01 19:45:30,924 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading FSImage in > 1701 msecs > 2011-01-01 19:45:30,945 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks = 0 > 2011-01-01 19:45:30,948 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid blocks = 0 > 2011-01-01 19:45:30,958 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of under-replicated b > locks = 0 > 2011-01-01 19:45:30,963 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of over-replicated b > locks = 0 > 2011-01-01 19:45:30,966 INFO org.apache.hadoop.hdfs.StateChange: STATE* Leaving safe mode after 1 secs. > 2011-01-01 19:45:30,971 INFO org.apache.hadoop.hdfs.StateChange: STATE* Network topology has 0 racks and 0 dat > anodes > 2011-01-01 19:45:30,973 INFO org.apache.hadoop.hdfs.StateChange: STATE* UnderReplicatedBlocks has 0 blocks > 2011-01-01 19:45:33,929 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) vi > a org.mortbay.log.Slf4jLog > 2011-01-01 19:45:35,020 INFO org.apache.hadoop.http.HttpServer: Port returned by webServer.getConnectors()[0]. > getLocalPort() before open() is -1. Opening the listener on 50070 > 2011-01-01 19:45:35,036 INFO org.apache.hadoop.http.HttpServer: listener.getLocalPort() returned 50070 webServ > er.getConnectors()[0].getLocalPort() returned 50070 > 2011-01-01 19:45:35,038 INFO org.apache.hadoop.http.HttpServer: Jetty bound to port 50070 > 2011-01-01 19:45:35,041 INFO org.mortbay.log: jetty-6.1.14 > sc-ssh-svr1 logs $ > > On Dec 31, 2010, at 4:28 PM, li ping wrote: > >> I suggest you should look through the logs to see if there is any error. >> And the second point that I need to point out is which node you run the >> command "hadoop fs -ls ". If you run the command on Node A, the >> configuration item "fs.default.name" should point to the HDFS. >> >> On Sat, Jan 1, 2011 at 3:20 AM, Jon Lederman <jon2...@gmail.com> wrote: >> >>> Hi Michael, >>> >>> Thanks for your response. It doesn't seem to be an issue with safemode. >>> >>> Even when I try the command dfsadmin -safemode get, the system hangs. I am >>> unable to execute any FS shell commands other than hadoop fs -help. >>> >>> I am wondering whether this an issue with communication between the >>> daemons? What should I be looking at there? Or could it be something else? >>> >>> When I do jps, I do see all the daemons listed. >>> >>> Any other thoughts. >>> >>> Thanks again and happy new year. >>> >>> -Jon >>> On Dec 31, 2010, at 9:09 AM, Black, Michael (IS) wrote: >>> >>>> Try checking your dfs status >>>> >>>> hadoop dfsadmin -safemode get >>>> >>>> Probably says "ON" >>>> >>>> hadoop dfsadmin -safemode leave >>>> >>>> Somebody else can probably say how to make this happen every reboot.... >>>> >>>> Michael D. Black >>>> Senior Scientist >>>> Advanced Analytics Directorate >>>> Northrop Grumman Information Systems >>>> >>>> >>>> ________________________________ >>>> >>>> From: Jon Lederman [mailto:jon2...@gmail.com] >>>> Sent: Fri 12/31/2010 11:00 AM >>>> To: common-user@hadoop.apache.org >>>> Subject: EXTERNAL:HDFS FS Commands Hanging System >>>> >>>> >>>> >>>> Hi All, >>>> >>>> I have been working on running Hadoop on a new microprocessor >>> architecture in pseudo-distributed mode. I have been successful in getting >>> SSH configured. I am also able to start a namenode, secondary namenode, >>> tasktracker, jobtracker and datanode as evidenced by the response I get from >>> jps. >>>> >>>> However, when I attempt to interact with the file system in any way such >>> as the simple command hadoop fs -ls, the system hangs. So it appears to me >>> that some communication is not occurring properly. Does anyone have any >>> suggestions what I look into in order to fix this problem? >>>> >>>> Thanks in advance. >>>> >>>> -Jon >>>> >>> >>> >> >> >> -- >> -----李平 >