Hi Dennis, No problem! :-) You did it really fast quite honestly. I will start the release process shortly...
Take care! Cheers, Chris On 4/2/07 6:21 PM, "Dennis Kubes" <[EMAIL PROTECTED]> wrote: > Chris, > > I have updated changes and resolved and closed the issue. Sorry about > not getting to it sooner. > > Dennis Kubes > > Chris Mattmann wrote: >> Hi Dennis, >> >> Thanks for taking care of this. :-) Could you update CHANGES.txt as well? >> Once you take care of that, in about 2 hrs (when I get home), I'll begin the >> release process again. >> >> Thanks! >> >> Cheers, >> Chris >> >> >> >> On 4/2/07 2:40 PM, "[EMAIL PROTECTED]" <[EMAIL PROTECTED]> wrote: >> >>> Author: kubes >>> Date: Mon Apr 2 14:40:10 2007 >>> New Revision: 524932 >>> >>> URL: http://svn.apache.org/viewvc?view=rev&rev=524932 >>> Log: >>> NUTCH-333 - SegmentMerger and SegmentReader should use NutchJob. Patch >>> supplied originally by Michael Stack and updated by Doğacan Güney. >>> >>> Modified: >>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java >>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java >>> >>> Modified: >>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java >>> URL: >>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/se >>> gm >>> ent/SegmentMerger.java?view=diff&rev=524932&r1=524931&r2=524932 >>> ============================================================================ >>> == >>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java >>> (original) >>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java >>> Mon Apr 2 14:40:10 2007 >>> @@ -18,17 +18,37 @@ >>> package org.apache.nutch.segment; >>> >>> import java.io.IOException; >>> -import java.util.*; >>> +import java.util.ArrayList; >>> +import java.util.HashMap; >>> +import java.util.Iterator; >>> +import java.util.TreeMap; >>> >>> import org.apache.commons.logging.Log; >>> import org.apache.commons.logging.LogFactory; >>> - >>> -import org.apache.hadoop.conf.*; >>> +import org.apache.hadoop.conf.Configuration; >>> +import org.apache.hadoop.conf.Configured; >>> import org.apache.hadoop.fs.FileSystem; >>> import org.apache.hadoop.fs.Path; >>> import org.apache.hadoop.fs.PathFilter; >>> -import org.apache.hadoop.io.*; >>> -import org.apache.hadoop.mapred.*; >>> +import org.apache.hadoop.io.MapFile; >>> +import org.apache.hadoop.io.SequenceFile; >>> +import org.apache.hadoop.io.Text; >>> +import org.apache.hadoop.io.UTF8; >>> +import org.apache.hadoop.io.Writable; >>> +import org.apache.hadoop.io.WritableComparable; >>> +import org.apache.hadoop.mapred.FileSplit; >>> +import org.apache.hadoop.mapred.InputSplit; >>> +import org.apache.hadoop.mapred.JobClient; >>> +import org.apache.hadoop.mapred.JobConf; >>> +import org.apache.hadoop.mapred.Mapper; >>> +import org.apache.hadoop.mapred.OutputCollector; >>> +import org.apache.hadoop.mapred.OutputFormatBase; >>> +import org.apache.hadoop.mapred.RecordReader; >>> +import org.apache.hadoop.mapred.RecordWriter; >>> +import org.apache.hadoop.mapred.Reducer; >>> +import org.apache.hadoop.mapred.Reporter; >>> +import org.apache.hadoop.mapred.SequenceFileInputFormat; >>> +import org.apache.hadoop.mapred.SequenceFileRecordReader; >>> import org.apache.hadoop.util.Progressable; >>> import org.apache.nutch.crawl.CrawlDatum; >>> import org.apache.nutch.crawl.Generator; >>> @@ -39,6 +59,7 @@ >>> import org.apache.nutch.parse.ParseText; >>> import org.apache.nutch.protocol.Content; >>> import org.apache.nutch.util.NutchConfiguration; >>> +import org.apache.nutch.util.NutchJob; >>> >>> /** >>> * This tool takes several segments and merges their data together. Only >>> the >>> @@ -482,7 +503,7 @@ >>> if (LOG.isInfoEnabled()) { >>> LOG.info("Merging " + segs.length + " segments to " + out + "/" + >>> segmentName); >>> } >>> - JobConf job = new JobConf(getConf()); >>> + JobConf job = new NutchJob(getConf()); >>> job.setJobName("mergesegs " + out + "/" + segmentName); >>> job.setBoolean("segment.merger.filter", filter); >>> job.setLong("segment.merger.slice", slice); >>> >>> Modified: >>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java >>> URL: >>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/se >>> gm >>> ent/SegmentReader.java?view=diff&rev=524932&r1=524931&r2=524932 >>> ============================================================================ >>> == >>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java >>> (original) >>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java >>> Mon Apr 2 14:40:10 2007 >>> @@ -17,18 +17,48 @@ >>> >>> package org.apache.nutch.segment; >>> >>> -import java.io.*; >>> +import java.io.BufferedReader; >>> +import java.io.BufferedWriter; >>> +import java.io.IOException; >>> +import java.io.InputStreamReader; >>> +import java.io.OutputStreamWriter; >>> +import java.io.PrintStream; >>> +import java.io.PrintWriter; >>> +import java.io.Writer; >>> import java.text.SimpleDateFormat; >>> -import java.util.*; >>> +import java.util.ArrayList; >>> +import java.util.Arrays; >>> +import java.util.Date; >>> +import java.util.HashMap; >>> +import java.util.Iterator; >>> +import java.util.List; >>> +import java.util.Map; >>> >>> import org.apache.commons.logging.Log; >>> import org.apache.commons.logging.LogFactory; >>> - >>> import org.apache.hadoop.conf.Configuration; >>> import org.apache.hadoop.conf.Configured; >>> -import org.apache.hadoop.fs.*; >>> -import org.apache.hadoop.io.*; >>> -import org.apache.hadoop.mapred.*; >>> +import org.apache.hadoop.fs.FileSystem; >>> +import org.apache.hadoop.fs.Path; >>> +import org.apache.hadoop.fs.PathFilter; >>> +import org.apache.hadoop.io.MapFile; >>> +import org.apache.hadoop.io.ObjectWritable; >>> +import org.apache.hadoop.io.SequenceFile; >>> +import org.apache.hadoop.io.Text; >>> +import org.apache.hadoop.io.UTF8; >>> +import org.apache.hadoop.io.Writable; >>> +import org.apache.hadoop.io.WritableComparable; >>> +import org.apache.hadoop.mapred.JobClient; >>> +import org.apache.hadoop.mapred.JobConf; >>> +import org.apache.hadoop.mapred.MapFileOutputFormat; >>> +import org.apache.hadoop.mapred.MapReduceBase; >>> +import org.apache.hadoop.mapred.Mapper; >>> +import org.apache.hadoop.mapred.OutputCollector; >>> +import org.apache.hadoop.mapred.RecordWriter; >>> +import org.apache.hadoop.mapred.Reducer; >>> +import org.apache.hadoop.mapred.Reporter; >>> +import org.apache.hadoop.mapred.SequenceFileInputFormat; >>> +import org.apache.hadoop.mapred.SequenceFileOutputFormat; >>> import org.apache.hadoop.util.Progressable; >>> import org.apache.nutch.crawl.CrawlDatum; >>> import org.apache.nutch.parse.ParseData; >>> @@ -36,6 +66,7 @@ >>> import org.apache.nutch.protocol.Content; >>> import org.apache.nutch.util.LogUtil; >>> import org.apache.nutch.util.NutchConfiguration; >>> +import org.apache.nutch.util.NutchJob; >>> >>> /** Dump the content of a segment. */ >>> public class SegmentReader extends Configured implements Reducer { >>> @@ -120,7 +151,7 @@ >>> } >>> >>> private JobConf createJobConf() { >>> - JobConf job = new JobConf(getConf()); >>> + JobConf job = new NutchJob(getConf()); >>> job.setBoolean("segment.reader.co", this.co); >>> job.setBoolean("segment.reader.fe", this.fe); >>> job.setBoolean("segment.reader.ge", this.ge); >>> >>> >> >>