Hi Tariq,
Thanks a lot for your help. Can you please let me know the path where I can check the old files in the HDFS and remove them accordingly. I am sorry to bother with these questions, I am absolutely new to Hadoop. Thanks again for your time and pateince. Regards, Indrashish On Tue, 8 Oct 2013 23:51:30 +0530, Mohammad Tariq wrote: > You don't have any more space left in your HDFS. Delete some old data or add additional storage. > > Warm Regards, > Tariq > cloudfront.blogspot.com [6] > > On Tue, Oct 8, 2013 at 11:47 PM, Basu,Indrashish wrote: > >> Hi , >> >> Just to update on this, I have deleted all the old logs and files from the /tmp and /app/hadoop directory, and restarted all the nodes, I have now 1 datanode available as per the below information : >> >> Configured Capacity: 3665985536 (3.41 GB) >> Present Capacity: 24576 (24 KB) >> >> DFS Remaining: 0 (0 KB) DFS Used: 24576 (24 KB) >> DFS Used%: 100% >> >> Under replicated blocks: 0 >> Blocks with corrupt replicas: 0 >> Missing blocks: 0 >> >> ------------------------------------------------- Datanodes available: 1 (1 total, 0 dead) >> >> Name: 10.227.56.195:50010 [5] >> Decommission Status : Normal >> Configured Capacity: 3665985536 (3.41 GB) >> DFS Used: 24576 (24 KB) >> Non DFS Used: 3665960960 (3.41 GB) >> DFS Remaining: 0(0 KB) >> DFS Used%: 0% >> DFS Remaining%: 0% >> Last contact: Tue Oct 08 11:12:19 PDT 2013 >> >> However when I tried putting the files back in HDFS, I am getting the same error as stated earlier. Do I need to clear some space for the HDFS ? >> >> Regards, >> Indrashish >> >> On Tue, 08 Oct 2013 14:01:19 -0400, Basu,Indrashish wrote: >> >>> Hi Jitendra, >>> >>> This is what I am getting in the datanode logs : >>> >>> 2013-10-07 11:27:41,960 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: Storage directory >>> /app/hadoop/tmp/dfs/data is not formatted. >>> 2013-10-07 11:27:41,961 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: Formatting ... >>> 2013-10-07 11:27:42,094 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: Registered >>> FSDatasetStatusMBean >>> 2013-10-07 11:27:42,099 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server at >>> 50010 >>> 2013-10-07 11:27:42,107 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is >>> 1048576 bytes/s >>> 2013-10-07 11:27:42,369 INFO org.mortbay.log: Logging to >>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>> org.mortbay.log.Slf4jLog >>> 2013-10-07 11:27:42,632 INFO org.apache.hadoop.http.HttpServer: Port >>> returned by webServer.getConnectors()[0].getLocalPort() before open() >>> is -1. Opening the listener on 50075 >>> 2013-10-07 11:27:42,633 INFO org.apache.hadoop.http.HttpServer: >>> listener.getLocalPort() returned 50075 >>> webServer.getConnectors()[0].getLocalPort() returned 50075 >>> 2013-10-07 11:27:42,634 INFO org.apache.hadoop.http.HttpServer: Jetty >>> bound to port 50075 >>> 2013-10-07 11:27:42,634 INFO org.mortbay.log: jetty-6.1.14 >>> 2013-10-07 11:31:29,821 INFO org.mortbay.log: Started >>> SelectChannelConnector@0.0.0.0:50075 [2] >>> 2013-10-07 11:31:29,843 INFO >>> org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics >>> with processName=DataNode, sessionId=null >>> 2013-10-07 11:31:29,912 INFO >>> org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics >>> with hostName=DataNode, port=50020 >>> 2013-10-07 11:31:29,922 INFO org.apache.hadoop.ipc.Server: IPC Server >>> Responder: starting >>> 2013-10-07 11:31:29,922 INFO org.apache.hadoop.ipc.Server: IPC Server >>> listener on 50020: starting >>> 2013-10-07 11:31:29,933 INFO org.apache.hadoop.ipc.Server: IPC Server >>> handler 0 on 50020: starting >>> 2013-10-07 11:31:29,933 INFO org.apache.hadoop.ipc.Server: IPC Server >>> handler 1 on 50020: starting >>> 2013-10-07 11:31:29,933 INFO org.apache.hadoop.ipc.Server: IPC Server >>> handler 2 on 50020: starting >>> 2013-10-07 11:31:29,934 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration = >>> DatanodeRegistration(tegra-ubuntu:50010, storageID=, infoPort=50075, >>> ipcPort=50020) >>> 2013-10-07 11:31:29,971 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: New storage id >>> DS-1027334635-127.0.1.1-50010-1381170689938 is assigned to data-node >>> 10.227.56.195:50010 [3] >>> 2013-10-07 11:31:29,973 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: >>> DatanodeRegistration(10.227.56.195:50010 [4], >>> storageID=DS-1027334635-127.0.1.1-50010-1381170689938, infoPort=50075, >>> ipcPort=50020)In DataNode.run, data = FSDataset >>> {dirpath='/app/hadoop/tmp/dfs/data/current'} >>> 2013-10-07 11:31:29,974 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: using >>> BLOCKREPORT_INTERVAL of 3600000msec Initial delay: 0msec >>> 2013-10-07 11:31:30,032 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 >>> blocks got processed in 19 msecs >>> 2013-10-07 11:31:30,035 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic >>> block scanner. >>> 2013-10-07 11:41:42,222 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 >>> blocks got processed in 20 msecs >>> 2013-10-07 12:41:43,482 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 >>> blocks got processed in 22 msecs >>> 2013-10-07 13:41:44,755 INFO >>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 >>> blocks got processed in 13 msecs >>> >>> I restarted the datanode and made sure that it is up and running >>> (typed jps command). >>> >>> Regards, >>> Indrashish >>> >>> On Tue, 8 Oct 2013 23:25:25 +0530, Jitendra Yadav wrote: >>> >>>> As per your dfs report, available DataNodes count is ZERO in you cluster. >>>> >>>> Please check your data node logs. >>>> >>>> Regards >>>> Jitendra >>>> >>>> On 10/8/13, Basu,Indrashish wrote: >>>> >>>>> Hello, >>>>> >>>>> My name is Indrashish Basu and I am a Masters student in the Department >>>>> of Electrical and Computer Engineering. >>>>> >>>>> Currently I am doing my research project on Hadoop implementation on >>>>> ARM processor and facing an issue while trying to run a sample Hadoop >>>>> source code on the same. Every time I am trying to put some files in the >>>>> HDFS, I am getting the below error. >>>>> >>>>> 13/10/07 11:31:29 WARN hdfs.DFSClient: DataStreamer Exception: >>>>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: File >>>>> /user/root/bin/cpu-kmeans2D could only be replicated to 0 nodes, instead >>>>> of 1 >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1267) >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at >>>>> >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>> at >>>>> >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) >>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >>>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >>>>> >>>>> at org.apache.hadoop.ipc.Client.call(Client.java:739) >>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >>>>> at com.sun.proxy.$Proxy0.addBlock(Unknown Source) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at >>>>> >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>> at >>>>> >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>>> at >>>>> >>>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) >>>>> at >>>>> >>>>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) >>>>> at com.sun.proxy.$Proxy0.addBlock(Unknown Source) >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.locateFollowingBlock(DFSClient.java:2904) >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2786) >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2076) >>>>> at >>>>> >>>>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2262) >>>>> >>>>> 13/10/07 11:31:29 WARN hdfs.DFSClient: Error Recovery for block null >>>>> bad datanode[0] nodes == null >>>>> 13/10/07 11:31:29 WARN hdfs.DFSClient: Could not get block locations. >>>>> Source file "/user/root/bin/cpu-kmeans2D" - Aborting... >>>>> put: java.io.IOException: File /user/root/bin/cpu-kmeans2D could only >>>>> be replicated to 0 nodes, instead of 1 >>>>> >>>>> I tried replicating the namenode and datanode by deleting all the old >>>>> logs on the master and the slave nodes as well as the folders under >>>>> /app/hadoop/, after which I formatted the namenode and started the >>>>> process again (bin/start-all.sh), but still no luck with the same. >>>>> >>>>> I tried generating the admin report(pasted below) after doing the >>>>> restart, it seems the data node is not getting started. >>>>> >>>>> ------------------------------------------------- >>>>> Datanodes available: 0 (0 total, 0 dead) >>>>> >>>>> root@tegra-ubuntu:~/hadoop-gpu-master/hadoop-gpu-0.20.1# bin/hadoop >>>>> dfsadmin -report >>>>> Configured Capacity: 0 (0 KB) >>>>> Present Capacity: 0 (0 KB) >>>>> DFS Remaining: 0 (0 KB) >>>>> DFS Used: 0 (0 KB) >>>>> DFS Used%: �% >>>>> Under replicated blocks: 0 >>>>> Blocks with corrupt replicas: 0 >>>>> Missing blocks: 0 >>>>> >>>>> ------------------------------------------------- >>>>> Datanodes available: 0 (0 total, 0 dead) >>>>> >>>>> I have tried the following methods to debug the process : >>>>> >>>>> 1) I logged in to the HADOOP home directory and removed all the old >>>>> logs (rm -rf logs/*) >>>>> >>>>> 2) Next I deleted the contents of the directory on all my slave and >>>>> master nodes (rm -rf /app/hadoop/*) >>>>> >>>>> 3) I formatted the namenode (bin/hadoop namenode -format) >>>>> >>>>> 4) I started all the processes - first the namenode, datanode and then >>>>> the map - reduce. I typed jps on the terminal to ensure that all the >>>>> processes (Namenode, Datanode, JobTracker, Task Tracker) are up and >>>>> running. >>>>> >>>>> 5) Now doing this, I recreated the directories in the dfs. >>>>> >>>>> However still no luck with the process. >>>>> >>>>> Can you kindly assist regarding this ? I am new to Hadoop and I am >>>>> having no idea as how I can proceed with this. >>>>> >>>>> Regards, >>>>> >>>>> -- >>>>> Indrashish Basu >>>>> Graduate Student >>>>> Department of Electrical and Computer Engineering >>>>> University of Florida >> >> -- >> Indrashish Basu >> Graduate Student >> Department of Electrical and Computer Engineering >> University of Florida -- Indrashish Basu Graduate Student Department of Electrical and Computer Engineering University of Florida Links: ------ [1] mailto:indrash...@ufl.edu [2] http://SelectChannelConnector@0.0.0.0:50075 [3] http://10.227.56.195:50010 [4] http://10.227.56.195:50010 [5] http://10.227.56.195:50010 [6] http://cloudfront.blogspot.com [7] mailto:indrash...@ufl.edu