Hi, I am running a map-reduce program on 6-Node ec2 cluster. and after a couple of hours all my tasks gets hanged.
so i started digging into the logs.... there were no logs for regionserver no logs for tasktracker. However for jobtracker i get the following: 2009-04-16 03:00:29,691 INFO org.apache.hadoop.ipc.Server: IPC Server handler 9 on 50002, call heartbeat(org.apache.hadoop.mapred.tasktrackersta...@2eed7d11, false, true, 10745) from 10.254.27.79:44222: error: java.io.IOException: org.apache.hadoop.fs.FSError: java.io.IOException: No space left on device java.io.IOException: org.apache.hadoop.fs.FSError: java.io.IOException: No space left on device at org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:199) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65) at java.io.BufferedOutputStream.write(BufferedOutputStream.java:109) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47) at java.io.DataOutputStream.write(DataOutputStream.java:90) at org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.writeChunk(ChecksumFileSystem.java:346) at org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:150) at org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:100) at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:86) at org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47) at java.io.DataOutputStream.write(DataOutputStream.java:90) at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:202) at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:297) at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:130) at java.io.OutputStreamWriter.close(OutputStreamWriter.java:216) at java.io.BufferedWriter.close(BufferedWriter.java:248) at java.io.PrintWriter.close(PrintWriter.java:295) at org.apache.hadoop.mapred.JobHistory$JobInfo.logFinished(JobHistory.java:1024) at org.apache.hadoop.mapred.JobInProgress.jobComplete(JobInProgress.java:1906) at org.apache.hadoop.mapred.JobInProgress.comp following are the disk information on dfs UI domU-12-31-39-00-0C-A1<http://domu-12-31-39-00-0c-a1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>0In Service413.380.8321.19391.360.2 94.672353 domU-12-31-39-00-16-F1<http://domu-12-31-39-00-16-f1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>1In Service413.380.4621.24391.670.11 94.752399 domU-12-31-39-00-45-71<http://domu-12-31-39-00-45-71.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>1In Service413.380.6421.34391.40.16 94.682303 domU-12-31-39-00-E5-D2<http://domu-12-31-39-00-e5-d2.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>0In Service413.380.6621.53391.180.16 94.632319 domU-12-31-39-01-64-12<http://domu-12-31-39-01-64-12.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>2In Service413.380.6421.24391.490.16 94.712264 domU-12-31-39-01-78-D1<http://domu-12-31-39-01-78-d1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F>0In Service413.380.4921.24391.650.12 94.741952 I m using hadoop 0.19.0 and hbase 0.19.0 n googling the error i came arcoss the JIRA issue http://issues.apache.org/jira/browse/HADOOP-4163 which says tht its been fixed in this version. :( Has anyone else come up with this exception? how do we check the maximum capacity for usable dfs and non usable dfs. Thanks Raakhi,