Thanks, I will check tht Regards, Raakhi
On Thu, Apr 16, 2009 at 1:42 PM, Miles Osborne <mi...@inf.ed.ac.uk> wrote: > it may be that intermediate results are filling your disks and when > the jobs crash, this all gets deleted. so it would look like you have > spare space when in reality you don't. > > i would check on the file system as your jobs run and see if indeed > they are filling-up. > > Miles > > 2009/4/16 Rakhi Khatwani <rakhi.khatw...@gmail.com>: > > Hi, > > following is the output on the df command > > [r...@domu-12-31-39-00-e5-d2 conf]# df -h > > Filesystem Size Used Avail Use% Mounted on > > /dev/sda1 9.9G 4.2G 5.2G 45% / > > /dev/sdb 414G 924M 392G 1% /mnt > > > > from the o/p it seems that i have quite an amount of memory available. > but i > > still get the exception :( > > > > Thanks > > Raakhi > > > > On Thu, Apr 16, 2009 at 1:18 PM, Desai, Milind B <milind.de...@hp.com > >wrote: > > > >> From the exception it appears that there is no space left on machine. > You > >> can check using 'df' > >> > >> Thanks > >> Milind > >> > >> -----Original Message----- > >> From: Rakhi Khatwani [mailto:rakhi.khatw...@gmail.com] > >> Sent: Thursday, April 16, 2009 1:15 PM > >> To: hbase-u...@hadoop.apache.org; core-user@hadoop.apache.org > >> Subject: No space left on device Exception > >> > >> Hi, > >> I am running a map-reduce program on 6-Node ec2 cluster. and after a > >> couple of hours all my tasks gets hanged. > >> > >> so i started digging into the logs.... > >> > >> there were no logs for regionserver > >> no logs for tasktracker. > >> However for jobtracker i get the following: > >> > >> 2009-04-16 03:00:29,691 INFO org.apache.hadoop.ipc.Server: IPC Server > >> handler 9 on 50002, call > >> heartbeat(org.apache.hadoop.mapred.tasktrackersta...@2eed7d11, false, > >> true, > >> 10745) from 10.254.27.79:44222: error: java.io.IOException: > >> org.apache.hadoop.fs.FSError: java.io.IOException: No space left on > device > >> java.io.IOException: org.apache.hadoop.fs.FSError: java.io.IOException: > No > >> space left on device > >> at > >> > >> > org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileOutputStream.write(RawLocalFileSystem.java:199) > >> at > >> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65) > >> at > java.io.BufferedOutputStream.write(BufferedOutputStream.java:109) > >> at > >> > >> > org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47) > >> at java.io.DataOutputStream.write(DataOutputStream.java:90) > >> at > >> > >> > org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSOutputSummer.writeChunk(ChecksumFileSystem.java:346) > >> at > >> > >> > org.apache.hadoop.fs.FSOutputSummer.writeChecksumChunk(FSOutputSummer.java:150) > >> at > >> org.apache.hadoop.fs.FSOutputSummer.write1(FSOutputSummer.java:100) > >> at > org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:86) > >> at > >> > >> > org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:47) > >> at java.io.DataOutputStream.write(DataOutputStream.java:90) > >> at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:202) > >> at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:297) > >> at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:130) > >> at java.io.OutputStreamWriter.close(OutputStreamWriter.java:216) > >> at java.io.BufferedWriter.close(BufferedWriter.java:248) > >> at java.io.PrintWriter.close(PrintWriter.java:295) > >> at > >> > >> > org.apache.hadoop.mapred.JobHistory$JobInfo.logFinished(JobHistory.java:1024) > >> at > >> > org.apache.hadoop.mapred.JobInProgress.jobComplete(JobInProgress.java:1906) > >> at org.apache.hadoop.mapred.JobInProgress.comp > >> > >> > >> > >> following are the disk information on dfs UI > >> domU-12-31-39-00-0C-A1< > >> > http://domu-12-31-39-00-0c-a1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >0In > >> Service413.380.8321.19391.360.2 > >> 94.672353 domU-12-31-39-00-16-F1< > >> > http://domu-12-31-39-00-16-f1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >1In > >> Service413.380.4621.24391.670.11 > >> 94.752399 domU-12-31-39-00-45-71< > >> > http://domu-12-31-39-00-45-71.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >1In > >> Service413.380.6421.34391.40.16 > >> 94.682303 domU-12-31-39-00-E5-D2< > >> > http://domu-12-31-39-00-e5-d2.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >0In > >> Service413.380.6621.53391.180.16 > >> 94.632319 domU-12-31-39-01-64-12< > >> > http://domu-12-31-39-01-64-12.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >2In > >> Service413.380.6421.24391.490.16 > >> 94.712264 domU-12-31-39-01-78-D1< > >> > http://domu-12-31-39-01-78-d1.compute-1.internal:50075/browseDirectory.jsp?namenodeInfoPort=50070&dir=%2F > >> >0In > >> Service413.380.4921.24391.650.12 > >> 94.741952 > >> > >> I m using hadoop 0.19.0 and hbase 0.19.0 > >> > >> n googling the error i came arcoss the JIRA issue > >> http://issues.apache.org/jira/browse/HADOOP-4163 > >> > >> which says tht its been fixed in this version. :( > >> > >> Has anyone else come up with this exception? > >> > >> how do we check the maximum capacity for usable dfs and non usable dfs. > >> Thanks > >> Raakhi, > >> > > > > > > -- > The University of Edinburgh is a charitable body, registered in > Scotland, with registration number SC005336. >