[appuser@localhost bin]$ ./hadoop dfs -ls Found 2 items drwxr-xr-x - appuser supergroup 0 2011-04-26 09:41 /user/appuser/input drwxr-xr-x - appuser supergroup 0 2011-04-26 09:42 /user/appuser/output [appuser@localhost bin]$ ./hadoop dfs -ls /usr/lemon/wordcount/input Found 2 items -rw-r--r-- 1 appuser supergroup 22 2011-04-26 12:16 /usr/lemon/wordcount/input/file01 -rw-r--r-- 1 appuser supergroup 28 2011-04-26 12:16 /usr/lemon/wordcount/input/file02 [appuser@localhost bin]$
On Fri, Jun 17, 2011 at 10:50 PM, Mostafa Gaber <moustafa.ga...@gmail.com>wrote: > Can you send us the output of "hadoop dfs -ls"? > > > > On Jun 17, 2011, at 10:21, Lemon Cheng <lemon...@gmail.com> wrote: > > Hi, > > Thanks for your reply. > I am not sure that. How can I prove that? > > I checked the localhost:50070, it shows 1 live node and 0 dead node. > And the log "hadoop-appuser-datanode-localhost.localdomain.log" shows: > > ************************************************************/ > 2011-06-17 19:59:38,658 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: > /************************************************************ > STARTUP_MSG: Starting DataNode > STARTUP_MSG: host = localhost.localdomain/127.0.0.1 > STARTUP_MSG: args = [] > STARTUP_MSG: version = 0.20.2 > STARTUP_MSG: build = > <https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20> > https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r > 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 > ************************************************************/ > 2011-06-17 19:59:46,738 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Registered > FSDatasetStatusMBean > 2011-06-17 19:59:46,749 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at 50010 > 2011-06-17 19:59:46,752 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is > 1048576 bytes/s > 2011-06-17 19:59:46,812 INFO org.mortbay.log: Logging to > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via > org.mortbay.log.Slf4jLog > 2011-06-17 19:59:46,870 INFO org.apache.hadoop.http.HttpServer: Port > returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. > Opening the listener on 50075 > 2011-06-17 19:59:46,871 INFO org.apache.hadoop.http.HttpServer: > listener.getLocalPort() returned 50075 > webServer.getConnectors()[0].getLocalPort() returned 50075 > 2011-06-17 19:59:46,871 INFO org.apache.hadoop.http.HttpServer: Jetty bound > to port 50075 > 2011-06-17 19:59:46,875 INFO org.mortbay.log: jetty-6.1.14 > 2011-06-17 20:01:45,702 INFO org.mortbay.log: Started > SelectChannelConnector@0.0.0.0:50075 > 2011-06-17 20:01:45,709 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=DataNode, sessionId=null > 2011-06-17 20:01:45,743 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: > Initializing RPC Metrics with hostName=DataNode, port=50020 > 2011-06-17 20:01:45,751 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration = > DatanodeRegistration(localhost.localdomain:50010, > storageID=DS-993704729-127.0.0.1-50010-1308296320968, infoPort=50075, > ipcPort=50020) > 2011-06-17 20:01:45,751 INFO org.apache.hadoop.ipc.Server: IPC Server > listener on 50020: starting > 2011-06-17 20:01:45,753 INFO org.apache.hadoop.ipc.Server: IPC Server > Responder: starting > 2011-06-17 20:01:45,754 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 2 on 50020: starting > 2011-06-17 20:01:45,754 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 0 on 50020: starting > 2011-06-17 20:01:45,754 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 1 on 50020: starting > 2011-06-17 20:01:45,795 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration( > 127.0.0.1:50010, storageID=DS-993704729-127.0.0.1-50010-1308296320968, > infoPort=50075, ipcPort=50020)In DataNode.run, data = > FSDataset{dirpath='/tmp/hadoop-appuser/dfs/data/current'} > 2011-06-17 20:01:45,799 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: using BLOCKREPORT_INTERVAL > of 3600000msec Initial delay: 0msec > 2011-06-17 20:01:45,828 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks got > processed in 11 msecs > 2011-06-17 20:01:45,833 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block > scanner. > 2011-06-17 20:56:02,945 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks got > processed in 1 msecs > 2011-06-17 21:56:02,248 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks got > processed in 1 msecs > > > On Fri, Jun 17, 2011 at 9:42 PM, Marcos Ortiz < <mlor...@uci.cu> > mlor...@uci.cu> wrote: > >> ** >> On 06/17/2011 07:41 AM, Lemon Cheng wrote: >> >> Hi, >> >> I am using the hadoop-0.20.2. After calling ./start-all.sh, i can type >> "hadoop dfs -ls". >> However, when i type "hadoop dfs -cat /usr/lemon/wordcount/input/file01", >> the error is shown as follow. >> I have searched the related problem in the web, but i can't find a >> solution for helping me to solve this problem. >> Anyone can give suggestion? >> Many Thanks. >> >> >> >> 11/06/17 19:27:12 INFO hdfs.DFSClient: No node available for block: >> blk_7095683278339921538_1029 file=/usr/lemon/wordcount/input/file01 >> 11/06/17 19:27:12 INFO hdfs.DFSClient: Could not obtain block >> blk_7095683278339921538_1029 from any node: java.io.IOException: No live >> nodes contain current block >> 11/06/17 19:27:15 INFO hdfs.DFSClient: No node available for block: >> blk_7095683278339921538_1029 file=/usr/lemon/wordcount/input/file01 >> 11/06/17 19:27:15 INFO hdfs.DFSClient: Could not obtain block >> blk_7095683278339921538_1029 from any node: java.io.IOException: No live >> nodes contain current block >> 11/06/17 19:27:18 INFO hdfs.DFSClient: No node available for block: >> blk_7095683278339921538_1029 file=/usr/lemon/wordcount/input/file01 >> 11/06/17 19:27:18 INFO hdfs.DFSClient: Could not obtain block >> blk_7095683278339921538_1029 from any node: java.io.IOException: No live >> nodes contain current block >> 11/06/17 19:27:21 WARN hdfs.DFSClient: DFS Read: java.io.IOException: >> Could not obtain block: blk_7095683278339921538_1029 >> file=/usr/lemon/wordcount/input/file01 >> at >> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.chooseDataNode(DFSClient.java:1812) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.blockSeekTo(DFSClient.java:1638) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.read(DFSClient.java:1767) >> at java.io.DataInputStream.read(DataInputStream.java:83) >> at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:47) >> at org.apache.hadoop.io.IOUtils.copyBytes(IOUtils.java:85) >> at org.apache.hadoop.fs.FsShell.printToStdout(FsShell.java:114) >> at org.apache.hadoop.fs.FsShell.access$100(FsShell.java:49) >> at org.apache.hadoop.fs.FsShell$1.process(FsShell.java:352) >> at >> org.apache.hadoop.fs.FsShell$DelayedExceptionThrowing.globAndProcess(FsShell.java:1898) >> at >> org.apache.hadoop.fs.FsShell.cat<http://org.apache.hadoop.fs.fsshell.cat/> >> (FsShell.java:346) >> at org.apache.hadoop.fs.FsShell.doall(FsShell.java:1543) >> at org.apache.hadoop.fs.FsShell.run(FsShell.java:1761) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) >> at org.apache.hadoop.fs.FsShell.main(FsShell.java:1880) >> >> >> Regards, >> Lemon >> >> Are you sure that all your DataNodes are online? >> >> >> -- >> Marcos Luís Ortíz Valmaseda >> Software Engineer (UCI) >> <http://marcosluis2186.posterous.com>http://marcosluis2186.posterous.com >> <http://twitter.com/marcosluis2186>http://twitter.com/marcosluis2186 >> >> >