Hi Dennis,

 No problem! :-) You did it really fast quite honestly. I will start the
release process shortly...

Take care!

Cheers,
  Chris





On 4/2/07 6:21 PM, "Dennis Kubes" <[EMAIL PROTECTED]> wrote:

> Chris,
> 
> I have updated changes and resolved and closed the issue.  Sorry about
> not getting to it sooner.
> 
> Dennis Kubes
> 
> Chris Mattmann wrote:
>> Hi Dennis,
>> 
>>  Thanks for taking care of this. :-) Could you update CHANGES.txt as well?
>> Once you take care of that, in about 2 hrs (when I get home), I'll begin the
>> release process again.
>> 
>> Thanks!
>> 
>> Cheers,
>>   Chris
>> 
>> 
>> 
>> On 4/2/07 2:40 PM, "[EMAIL PROTECTED]" <[EMAIL PROTECTED]> wrote:
>> 
>>> Author: kubes
>>> Date: Mon Apr  2 14:40:10 2007
>>> New Revision: 524932
>>> 
>>> URL: http://svn.apache.org/viewvc?view=rev&rev=524932
>>> Log:
>>> NUTCH-333 - SegmentMerger and SegmentReader should use NutchJob.  Patch
>>> supplied originally by Michael Stack and updated by Doğacan Güney.
>>> 
>>> Modified:
>>>     lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>>>     lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>>> 
>>> Modified: 
>>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>>> URL: 
>>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/se
>>> gm
>>> ent/SegmentMerger.java?view=diff&rev=524932&r1=524931&r2=524932
>>> ============================================================================
>>> ==
>>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>>> (original)
>>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>>> Mon Apr  2 14:40:10 2007
>>> @@ -18,17 +18,37 @@
>>>  package org.apache.nutch.segment;
>>>  
>>>  import java.io.IOException;
>>> -import java.util.*;
>>> +import java.util.ArrayList;
>>> +import java.util.HashMap;
>>> +import java.util.Iterator;
>>> +import java.util.TreeMap;
>>>  
>>>  import org.apache.commons.logging.Log;
>>>  import org.apache.commons.logging.LogFactory;
>>> -
>>> -import org.apache.hadoop.conf.*;
>>> +import org.apache.hadoop.conf.Configuration;
>>> +import org.apache.hadoop.conf.Configured;
>>>  import org.apache.hadoop.fs.FileSystem;
>>>  import org.apache.hadoop.fs.Path;
>>>  import org.apache.hadoop.fs.PathFilter;
>>> -import org.apache.hadoop.io.*;
>>> -import org.apache.hadoop.mapred.*;
>>> +import org.apache.hadoop.io.MapFile;
>>> +import org.apache.hadoop.io.SequenceFile;
>>> +import org.apache.hadoop.io.Text;
>>> +import org.apache.hadoop.io.UTF8;
>>> +import org.apache.hadoop.io.Writable;
>>> +import org.apache.hadoop.io.WritableComparable;
>>> +import org.apache.hadoop.mapred.FileSplit;
>>> +import org.apache.hadoop.mapred.InputSplit;
>>> +import org.apache.hadoop.mapred.JobClient;
>>> +import org.apache.hadoop.mapred.JobConf;
>>> +import org.apache.hadoop.mapred.Mapper;
>>> +import org.apache.hadoop.mapred.OutputCollector;
>>> +import org.apache.hadoop.mapred.OutputFormatBase;
>>> +import org.apache.hadoop.mapred.RecordReader;
>>> +import org.apache.hadoop.mapred.RecordWriter;
>>> +import org.apache.hadoop.mapred.Reducer;
>>> +import org.apache.hadoop.mapred.Reporter;
>>> +import org.apache.hadoop.mapred.SequenceFileInputFormat;
>>> +import org.apache.hadoop.mapred.SequenceFileRecordReader;
>>>  import org.apache.hadoop.util.Progressable;
>>>  import org.apache.nutch.crawl.CrawlDatum;
>>>  import org.apache.nutch.crawl.Generator;
>>> @@ -39,6 +59,7 @@
>>>  import org.apache.nutch.parse.ParseText;
>>>  import org.apache.nutch.protocol.Content;
>>>  import org.apache.nutch.util.NutchConfiguration;
>>> +import org.apache.nutch.util.NutchJob;
>>>  
>>>  /**
>>>   * This tool takes several segments and merges their data together. Only
>>> the
>>> @@ -482,7 +503,7 @@
>>>      if (LOG.isInfoEnabled()) {
>>>        LOG.info("Merging " + segs.length + " segments to " + out + "/" +
>>> segmentName);
>>>      }
>>> -    JobConf job = new JobConf(getConf());
>>> +    JobConf job = new NutchJob(getConf());
>>>      job.setJobName("mergesegs " + out + "/" + segmentName);
>>>      job.setBoolean("segment.merger.filter", filter);
>>>      job.setLong("segment.merger.slice", slice);
>>> 
>>> Modified: 
>>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>>> URL: 
>>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/se
>>> gm
>>> ent/SegmentReader.java?view=diff&rev=524932&r1=524931&r2=524932
>>> ============================================================================
>>> ==
>>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>>> (original)
>>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>>> Mon Apr  2 14:40:10 2007
>>> @@ -17,18 +17,48 @@
>>>  
>>>  package org.apache.nutch.segment;
>>>  
>>> -import java.io.*;
>>> +import java.io.BufferedReader;
>>> +import java.io.BufferedWriter;
>>> +import java.io.IOException;
>>> +import java.io.InputStreamReader;
>>> +import java.io.OutputStreamWriter;
>>> +import java.io.PrintStream;
>>> +import java.io.PrintWriter;
>>> +import java.io.Writer;
>>>  import java.text.SimpleDateFormat;
>>> -import java.util.*;
>>> +import java.util.ArrayList;
>>> +import java.util.Arrays;
>>> +import java.util.Date;
>>> +import java.util.HashMap;
>>> +import java.util.Iterator;
>>> +import java.util.List;
>>> +import java.util.Map;
>>>  
>>>  import org.apache.commons.logging.Log;
>>>  import org.apache.commons.logging.LogFactory;
>>> -
>>>  import org.apache.hadoop.conf.Configuration;
>>>  import org.apache.hadoop.conf.Configured;
>>> -import org.apache.hadoop.fs.*;
>>> -import org.apache.hadoop.io.*;
>>> -import org.apache.hadoop.mapred.*;
>>> +import org.apache.hadoop.fs.FileSystem;
>>> +import org.apache.hadoop.fs.Path;
>>> +import org.apache.hadoop.fs.PathFilter;
>>> +import org.apache.hadoop.io.MapFile;
>>> +import org.apache.hadoop.io.ObjectWritable;
>>> +import org.apache.hadoop.io.SequenceFile;
>>> +import org.apache.hadoop.io.Text;
>>> +import org.apache.hadoop.io.UTF8;
>>> +import org.apache.hadoop.io.Writable;
>>> +import org.apache.hadoop.io.WritableComparable;
>>> +import org.apache.hadoop.mapred.JobClient;
>>> +import org.apache.hadoop.mapred.JobConf;
>>> +import org.apache.hadoop.mapred.MapFileOutputFormat;
>>> +import org.apache.hadoop.mapred.MapReduceBase;
>>> +import org.apache.hadoop.mapred.Mapper;
>>> +import org.apache.hadoop.mapred.OutputCollector;
>>> +import org.apache.hadoop.mapred.RecordWriter;
>>> +import org.apache.hadoop.mapred.Reducer;
>>> +import org.apache.hadoop.mapred.Reporter;
>>> +import org.apache.hadoop.mapred.SequenceFileInputFormat;
>>> +import org.apache.hadoop.mapred.SequenceFileOutputFormat;
>>>  import org.apache.hadoop.util.Progressable;
>>>  import org.apache.nutch.crawl.CrawlDatum;
>>>  import org.apache.nutch.parse.ParseData;
>>> @@ -36,6 +66,7 @@
>>>  import org.apache.nutch.protocol.Content;
>>>  import org.apache.nutch.util.LogUtil;
>>>  import org.apache.nutch.util.NutchConfiguration;
>>> +import org.apache.nutch.util.NutchJob;
>>>  
>>>  /** Dump the content of a segment. */
>>>  public class SegmentReader extends Configured implements Reducer {
>>> @@ -120,7 +151,7 @@
>>>    }
>>>  
>>>    private JobConf createJobConf() {
>>> -    JobConf job = new JobConf(getConf());
>>> +    JobConf job = new NutchJob(getConf());
>>>      job.setBoolean("segment.reader.co", this.co);
>>>      job.setBoolean("segment.reader.fe", this.fe);
>>>      job.setBoolean("segment.reader.ge", this.ge);
>>> 
>>> 
>> 
>> 


Reply via email to