Hi,
i want to run nutch crawler command from tomcat web application with java
ProcessBuilder, when i run crawler command from terminal every thing is ok,
but when run with ProccessBuilder job fails with below error.
nutch version: 1.12
java version: 8
OS: ubuntu 16.04
tomcat version: 8
solr version: 6.2
thanks for your help.
java.lang.Exception: java.io.IOException: Mkdirs failed to create
file:/generate-temp-8dca91a6-3610-4802-a534-c0cdf85cde73/_temporary/0/_temporary/attempt_local1345668275_0001_r_000001_0/fetchlist-1
    at
org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462)
    at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529)
Caused by: java.io.IOException: Mkdirs failed to create
file:/generate-temp-8dca91a6-3610-4802-a534-c0cdf85cde73/_temporary/0/_temporary/attempt_local1345668275_0001_r_000001_0/fetchlist-1
    at
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:438)
    at
org.apache.hadoop.fs.ChecksumFileSystem.create(ChecksumFileSystem.java:424)
    at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906)
    at
org.apache.hadoop.io.SequenceFile$Writer.<init>(SequenceFile.java:1071)
    at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:270)
    at org.apache.hadoop.io.SequenceFile.createWriter(SequenceFile.java:527)
    at
org.apache.hadoop.mapred.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:63)
    at
org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat.getBaseRecordWriter(MultipleSequenceFileOutputFormat.java:51)
    at
org.apache.hadoop.mapred.lib.MultipleOutputFormat$1.write(MultipleOutputFormat.java:104)
    at
org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write(ReduceTask.java:493)
    at org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:422)
    at org.apache.nutch.crawl.Generator$Selector.reduce(Generator.java:342)
    at org.apache.nutch.crawl.Generator$Selector.reduce(Generator.java:110)
    at
org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444)
    at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392)
    at
org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319)
    at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)
2017-08-22 20:58:28,674 INFO  mapreduce.Job - Job job_local1345668275_0001
running in uber mode : false
2017-08-22 20:58:28,674 INFO  mapreduce.Job -  map 100% reduce 50%
2017-08-22 20:58:28,675 INFO  mapreduce.Job - Job job_local1345668275_0001
failed with state FAILED due to: NA
2017-08-22 20:58:28,684 INFO  mapreduce.Job - Counters: 33
    File System Counters
        FILE: Number of bytes read=1601520
        FILE: Number of bytes written=2209492
        FILE: Number of read operations=0
        FILE: Number of large read operations=0
        FILE: Number of write operations=0
    Map-Reduce Framework
        Map input records=1
        Map output records=1
        Map output bytes=77
        Map output materialized bytes=88
        Input split bytes=142
        Combine input records=0
        Combine output records=0
        Reduce input groups=0
        Reduce shuffle bytes=88
        Reduce input records=0
        Reduce output records=0
        Spilled Records=1
        Shuffled Maps =2
        Failed Shuffles=0
        Merged Map outputs=2
        GC time elapsed (ms)=6
        CPU time spent (ms)=0
        Physical memory (bytes) snapshot=0
        Virtual memory (bytes) snapshot=0
        Total committed heap usage (bytes)=531628032
    Shuffle Errors
        BAD_ID=0
        CONNECTION=0
        IO_ERROR=0
        WRONG_LENGTH=0
        WRONG_MAP=0
        WRONG_REDUCE=0
    File Input Format Counters
        Bytes Read=154
    File Output Format Counters
        Bytes Written=0
2017-08-22 20:58:28,684 ERROR crawl.Generator - Generator:
java.io.IOException: Job failed!
    at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836)
    at org.apache.nutch.crawl.Generator.generate(Generator.java:589)
    at org.apache.nutch.crawl.Generator.run(Generator.java:764)
    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
    at org.apache.nutch.crawl.Generator.main(Generator.java:717)

Reply via email to