Hi Dennis, Thanks for taking care of this. :-) Could you update CHANGES.txt as well? Once you take care of that, in about 2 hrs (when I get home), I'll begin the release process again.
Thanks! Cheers, Chris On 4/2/07 2:40 PM, "[EMAIL PROTECTED]" <[EMAIL PROTECTED]> wrote: > Author: kubes > Date: Mon Apr 2 14:40:10 2007 > New Revision: 524932 > > URL: http://svn.apache.org/viewvc?view=rev&rev=524932 > Log: > NUTCH-333 - SegmentMerger and SegmentReader should use NutchJob. Patch > supplied originally by Michael Stack and updated by Doğacan Güney. > > Modified: > lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java > lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java > > Modified: > lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java > URL: > http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/segm > ent/SegmentMerger.java?view=diff&rev=524932&r1=524931&r2=524932 > ============================================================================== > --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java > (original) > +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java > Mon Apr 2 14:40:10 2007 > @@ -18,17 +18,37 @@ > package org.apache.nutch.segment; > > import java.io.IOException; > -import java.util.*; > +import java.util.ArrayList; > +import java.util.HashMap; > +import java.util.Iterator; > +import java.util.TreeMap; > > import org.apache.commons.logging.Log; > import org.apache.commons.logging.LogFactory; > - > -import org.apache.hadoop.conf.*; > +import org.apache.hadoop.conf.Configuration; > +import org.apache.hadoop.conf.Configured; > import org.apache.hadoop.fs.FileSystem; > import org.apache.hadoop.fs.Path; > import org.apache.hadoop.fs.PathFilter; > -import org.apache.hadoop.io.*; > -import org.apache.hadoop.mapred.*; > +import org.apache.hadoop.io.MapFile; > +import org.apache.hadoop.io.SequenceFile; > +import org.apache.hadoop.io.Text; > +import org.apache.hadoop.io.UTF8; > +import org.apache.hadoop.io.Writable; > +import org.apache.hadoop.io.WritableComparable; > +import org.apache.hadoop.mapred.FileSplit; > +import org.apache.hadoop.mapred.InputSplit; > +import org.apache.hadoop.mapred.JobClient; > +import org.apache.hadoop.mapred.JobConf; > +import org.apache.hadoop.mapred.Mapper; > +import org.apache.hadoop.mapred.OutputCollector; > +import org.apache.hadoop.mapred.OutputFormatBase; > +import org.apache.hadoop.mapred.RecordReader; > +import org.apache.hadoop.mapred.RecordWriter; > +import org.apache.hadoop.mapred.Reducer; > +import org.apache.hadoop.mapred.Reporter; > +import org.apache.hadoop.mapred.SequenceFileInputFormat; > +import org.apache.hadoop.mapred.SequenceFileRecordReader; > import org.apache.hadoop.util.Progressable; > import org.apache.nutch.crawl.CrawlDatum; > import org.apache.nutch.crawl.Generator; > @@ -39,6 +59,7 @@ > import org.apache.nutch.parse.ParseText; > import org.apache.nutch.protocol.Content; > import org.apache.nutch.util.NutchConfiguration; > +import org.apache.nutch.util.NutchJob; > > /** > * This tool takes several segments and merges their data together. Only the > @@ -482,7 +503,7 @@ > if (LOG.isInfoEnabled()) { > LOG.info("Merging " + segs.length + " segments to " + out + "/" + > segmentName); > } > - JobConf job = new JobConf(getConf()); > + JobConf job = new NutchJob(getConf()); > job.setJobName("mergesegs " + out + "/" + segmentName); > job.setBoolean("segment.merger.filter", filter); > job.setLong("segment.merger.slice", slice); > > Modified: > lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java > URL: > http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/segm > ent/SegmentReader.java?view=diff&rev=524932&r1=524931&r2=524932 > ============================================================================== > --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java > (original) > +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java > Mon Apr 2 14:40:10 2007 > @@ -17,18 +17,48 @@ > > package org.apache.nutch.segment; > > -import java.io.*; > +import java.io.BufferedReader; > +import java.io.BufferedWriter; > +import java.io.IOException; > +import java.io.InputStreamReader; > +import java.io.OutputStreamWriter; > +import java.io.PrintStream; > +import java.io.PrintWriter; > +import java.io.Writer; > import java.text.SimpleDateFormat; > -import java.util.*; > +import java.util.ArrayList; > +import java.util.Arrays; > +import java.util.Date; > +import java.util.HashMap; > +import java.util.Iterator; > +import java.util.List; > +import java.util.Map; > > import org.apache.commons.logging.Log; > import org.apache.commons.logging.LogFactory; > - > import org.apache.hadoop.conf.Configuration; > import org.apache.hadoop.conf.Configured; > -import org.apache.hadoop.fs.*; > -import org.apache.hadoop.io.*; > -import org.apache.hadoop.mapred.*; > +import org.apache.hadoop.fs.FileSystem; > +import org.apache.hadoop.fs.Path; > +import org.apache.hadoop.fs.PathFilter; > +import org.apache.hadoop.io.MapFile; > +import org.apache.hadoop.io.ObjectWritable; > +import org.apache.hadoop.io.SequenceFile; > +import org.apache.hadoop.io.Text; > +import org.apache.hadoop.io.UTF8; > +import org.apache.hadoop.io.Writable; > +import org.apache.hadoop.io.WritableComparable; > +import org.apache.hadoop.mapred.JobClient; > +import org.apache.hadoop.mapred.JobConf; > +import org.apache.hadoop.mapred.MapFileOutputFormat; > +import org.apache.hadoop.mapred.MapReduceBase; > +import org.apache.hadoop.mapred.Mapper; > +import org.apache.hadoop.mapred.OutputCollector; > +import org.apache.hadoop.mapred.RecordWriter; > +import org.apache.hadoop.mapred.Reducer; > +import org.apache.hadoop.mapred.Reporter; > +import org.apache.hadoop.mapred.SequenceFileInputFormat; > +import org.apache.hadoop.mapred.SequenceFileOutputFormat; > import org.apache.hadoop.util.Progressable; > import org.apache.nutch.crawl.CrawlDatum; > import org.apache.nutch.parse.ParseData; > @@ -36,6 +66,7 @@ > import org.apache.nutch.protocol.Content; > import org.apache.nutch.util.LogUtil; > import org.apache.nutch.util.NutchConfiguration; > +import org.apache.nutch.util.NutchJob; > > /** Dump the content of a segment. */ > public class SegmentReader extends Configured implements Reducer { > @@ -120,7 +151,7 @@ > } > > private JobConf createJobConf() { > - JobConf job = new JobConf(getConf()); > + JobConf job = new NutchJob(getConf()); > job.setBoolean("segment.reader.co", this.co); > job.setBoolean("segment.reader.fe", this.fe); > job.setBoolean("segment.reader.ge", this.ge); > >