Your job tracker out-of-memory problem may be related to https://issues.apache.org/jira/browse/HADOOP-4766
Runping On Mon, Mar 2, 2009 at 4:29 PM, bzheng <bing.zh...@gmail.com> wrote: > > Thanks for all the info. Upon further investigation, we are dealing with > two > separate issues: > > 1. problem processing a lot of gz files > > we have tried the hadoop.native.lib setting and it makes little difference. > however, this is not that big a deal since we can use multiple jobs each > processing a small chunk of the files instead of one big job processing all > the files. > > 2. jobtracker out of memory > > by increasing amount of memory for the jobtracker, we can delay the > inevitable. since the jobtraker's memory usage keeps going up as we run > more job, we will need to restart the cluster once this error happens. we > are currently using 0.18.3 and are holding off changing to a different > version because we don't want to lose the existing files on HDFS. > > > bzheng wrote: > > > > I have about 24k gz files (about 550GB total) on hdfs and has a really > > simple java program to convert them into sequence files. If the script's > > setInputPaths takes a Path[] of all 24k files, it will get a OutOfMemory > > error at about 35% map complete. If I make the script process 2k files > > per job and run 12 jobs consecutively, then it goes through all files > > fine. The cluster I'm using has about 67 nodes. Each nodes has 16GB > > memory, max 7 map, and max 2 reduce. > > > > The map task is really simple, it takes LongWritable as key and Text as > > value, generate a Text newKey, and output.collect(Text newKey, Text > > value). It doesn't have any code that can possibly leak memory. > > > > There's no stack trace for the vast majority of the OutOfMemory error, > > there's just a single line in the log like this: > > > > 2009-02-23 14:27:50,902 INFO org.apache.hadoop.mapred.TaskTracker: > > java.lang.OutOfMemoryError: Java heap space > > > > I can't find the stack trace right now, but rarely the OutOfMemory error > > originates from some hadoop config array copy opertaion. There's no > > special config for the script. > > > > -- > View this message in context: > http://www.nabble.com/OutOfMemory-error-processing-large-amounts-of-gz-files-tp22193552p22300192.html > Sent from the Hadoop core-user mailing list archive at Nabble.com. > >