seems this was a file permissions error; deleting files generated by
hadoop in /tmp seems to have taken care of the the Disk error; not sure if
- this is the best thing to do?

but now looks like there is sudden thread death; no explanation:

2009-11-04 14:56:41,613 WARN  [org.apache.hadoop.mapred.LocalJobRunner]
job_local_0001
java.lang.ThreadDeath
        at java.lang.Thread.stop(Thread.java:715)
        at 
org.apache.hadoop.mapred.LocalJobRunner.killJob(LocalJobRunner.java:310)
        at
org.apache.hadoop.mapred.JobClient$NetworkedJob.killJob(JobClient.java:315)
        at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1239)
        at org.apache.nutch.segment.SegmentMerger.merge(SegmentMerger.java:620)
        at org.apache.nutch.segment.SegmentMerger.main(SegmentMerger.java:665)


> hi there,
>
> there was a little improvement; at least its not running out of ram
> anymore; but you're right there seems to be a side effect.
>
> i am now having what seems to be disk issues! i am running in a VPS so i
> am suspecting this might have something to do with it?
>
> but what is the cause now?
>
>
> ==>>
>
> 00:36:28,912 INFO [TaskRunner] Task 'attempt_local_0001_m_000064_0'
> done.
> 00:36:29,104 INFO  [MapTask] numReduceTasks: 1
> 00:36:29,104 INFO  [MapTask] io.sort.mb = 100
> 00:36:29,240 INFO  [MapTask] data buffer = 79691776/99614720
> 00:36:29,240 INFO  [MapTask] record buffer = 262144/327680
> 00:36:29,260 INFO  [CodecPool] Got brand-new decompressor
> 00:36:29,264 INFO  [MapTask] Starting flush of map output
> 00:36:29,276 INFO  [MapTask] Finished spill 0
> 00:36:29,280 INFO  [TaskRunner] Task:attempt_local_0001_m_000065_0 is
> done. And is in the process of commiting
> 00:36:29,280 INFO  [LocalJobRunner]
> file:/home/meda/workspace/web/crawl/segments/20091101171338/parse_text/part-00000/data:0+12655
> 00:36:29,280 INFO  [TaskRunner] Task 'attempt_local_0001_m_000065_0' done.
> 00:36:38,533 WARN  [LocalJobRunner] job_local_0001
> org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find
> taskTracker/jobcache/job_local_0001/attempt_local_0001_m_000000_0/output/file.out
> in any of the configured local directories
>       at
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathToRead(LocalDirAllocator.java:381)
>       at
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAllocator.java:138)
>       at
> org.apache.hadoop.mapred.MapOutputFile.getOutputFile(MapOutputFile.java:50)
>       at
> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:150)
> Exception in thread "main" java.io.IOException: Job failed!
>       at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232)
>       at org.apache.nutch.segment.SegmentMerger.merge(SegmentMerger.java:620)
>       at org.apache.nutch.segment.SegmentMerger.main(SegmentMerger.java:665)
>
>
>
> On Tue, 2009-11-03 at 10:28 -0500, Kalaimathan Mahenthiran wrote:
>> if you set the mapred.child.java.opts
>> with additional value "-XX: -UseGCOverheadLimit" you can bypass this
>> exception. I don't know if it has any side effects as a result of
>> this..
>> ex.
>> -Xmx512m -XX: -UseGCOverheadLimit
>>
>>
>> On Tue, Nov 3, 2009 at 7:50 AM, Fadzi Ushewokunze
>> <fa...@butterflycluster.net> wrote:
>> > hi,
>> >
>> > i am running on a single machine; 2G RAM, and java heap space set at
>> > 1024m, the segments are quite - tiny less than 100 urls and during
>> > mergeSegments i get this exception below;
>> >
>> > i have set mapred.child.java.opts=-Xmx512m but there is no change;
>> >
>> > any suggestions?
>> >
>> >
>> > ====>
>> >
>> > 2009-11-03 17:58:28,971 INFO
>> [org.apache.hadoop.mapred.LocalJobRunner]
>> > reduce > reduce
>> > 2009-11-03 17:58:38,448 INFO
>> [org.apache.hadoop.mapred.LocalJobRunner]
>> > reduce > reduce
>> > 2009-11-03 17:58:57,085 INFO
>> [org.apache.hadoop.mapred.LocalJobRunner]
>> > reduce > reduce
>> > 2009-11-03 17:59:34,723 INFO
>> [org.apache.hadoop.mapred.LocalJobRunner]
>> > reduce > reduce
>> > 2009-11-03 18:02:09,660 INFO  [org.apache.hadoop.mapred.TaskRunner]
>> > Communication exception: java.lang.OutOfMemoryError: Java heap space
>> >        at org.apache.hadoop.mapred.Counters
>> > $Group.getCounterForName(Counters.java:327)
>> >        at
>> > org.apache.hadoop.mapred.Counters.incrAllCounters(Counters.java:494)
>> >        at org.apache.hadoop.mapred.Counters.sum(Counters.java:506)
>> >        at org.apache.hadoop.mapred.LocalJobRunner
>> > $Job.statusUpdate(LocalJobRunner.java:222)
>> >        at org.apache.hadoop.mapred.Task$1.run(Task.java:418)
>> >        at java.lang.Thread.run(Thread.java:619)
>> >
>> > 2009-11-03 18:02:10,376 WARN
>> [org.apache.hadoop.mapred.LocalJobRunner]
>> > job_local_0001
>> > java.lang.ThreadDeath
>> >        at java.lang.Thread.stop(Thread.java:715)
>> >        at
>> > org.apache.hadoop.mapred.LocalJobRunner.killJob(LocalJobRunner.java:310)
>> >        at org.apache.hadoop.mapred.JobClient
>> > $NetworkedJob.killJob(JobClient.java:315)
>> >        at
>> > org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1239)
>> >        at
>> > org.apache.nutch.segment.SegmentMerger.merge(SegmentMerger.java:620)
>> >        at
>> > org.apache.nutch.segment.SegmentMerger.main(SegmentMerger.java:665)
>> >
>> >
>
>


Reply via email to