Sounds like mapred.task.timeout? The default is 10 minutes. http://hadoop.apache.org/common/docs/current/mapred-default.html
Thanks, Tom On Wed, Jan 18, 2012 at 2:05 PM, Steve Lewis <lordjoe2...@gmail.com> wrote: > The map tasks fail timing out after 600 sec. > I am processing one 9 GB file with 16,000,000 records. Each record (think > is it as a line) generates hundreds of key value pairs. > The job is unusual in that the output of the mapper in terms of records or > bytes orders of magnitude larger than the input. > I have no idea what is slowing down the job except that the problem is in > the writes. > > If I change the job to merely bypass a fraction of the context.write > statements the job succeeds. > This is one map task that failed and one that succeeded - I cannot > understand how a write can take so long > or what else the mapper might be doing > > JOB FAILED WITH TIMEOUT > > *Parser*TotalProteins90,103NumberFragments10,933,089 > *FileSystemCounters*HDFS_BYTES_READ67,245,605FILE_BYTES_WRITTEN444,054,807 > *Map-Reduce Framework*Combine output records10,033,499Map input records > 90,103Spilled Records10,032,836Map output bytes3,520,182,794Combine input > records10,844,881Map output records10,933,089 > Same code but fewer writes > JOB SUCCEEDED > > *Parser*TotalProteins90,103NumberFragments206,658,758 > *FileSystemCounters*FILE_BYTES_READ111,578,253HDFS_BYTES_READ67,245,607 > FILE_BYTES_WRITTEN220,169,922 > *Map-Reduce Framework*Combine output records4,046,128Map input > records90,103Spilled > Records4,046,128Map output bytes662,354,413Combine input records4,098,609Map > output records2,066,588 > Any bright ideas > -- > Steven M. Lewis PhD > 4221 105th Ave NE > Kirkland, WA 98033 > 206-384-1340 (cell) > Skype lordjoe_com