Chris,

I have updated changes and resolved and closed the issue.  Sorry about 
not getting to it sooner.

Dennis Kubes

Chris Mattmann wrote:
> Hi Dennis,
> 
>  Thanks for taking care of this. :-) Could you update CHANGES.txt as well?
> Once you take care of that, in about 2 hrs (when I get home), I'll begin the
> release process again.
> 
> Thanks!
> 
> Cheers,
>   Chris
> 
> 
> 
> On 4/2/07 2:40 PM, "[EMAIL PROTECTED]" <[EMAIL PROTECTED]> wrote:
> 
>> Author: kubes
>> Date: Mon Apr  2 14:40:10 2007
>> New Revision: 524932
>>
>> URL: http://svn.apache.org/viewvc?view=rev&rev=524932
>> Log:
>> NUTCH-333 - SegmentMerger and SegmentReader should use NutchJob.  Patch
>> supplied originally by Michael Stack and updated by Doğacan Güney.
>>
>> Modified:
>>     lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>>     lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>>
>> Modified: 
>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>> URL: 
>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/segm
>> ent/SegmentMerger.java?view=diff&rev=524932&r1=524931&r2=524932
>> ==============================================================================
>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>> (original)
>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentMerger.java
>> Mon Apr  2 14:40:10 2007
>> @@ -18,17 +18,37 @@
>>  package org.apache.nutch.segment;
>>  
>>  import java.io.IOException;
>> -import java.util.*;
>> +import java.util.ArrayList;
>> +import java.util.HashMap;
>> +import java.util.Iterator;
>> +import java.util.TreeMap;
>>  
>>  import org.apache.commons.logging.Log;
>>  import org.apache.commons.logging.LogFactory;
>> -
>> -import org.apache.hadoop.conf.*;
>> +import org.apache.hadoop.conf.Configuration;
>> +import org.apache.hadoop.conf.Configured;
>>  import org.apache.hadoop.fs.FileSystem;
>>  import org.apache.hadoop.fs.Path;
>>  import org.apache.hadoop.fs.PathFilter;
>> -import org.apache.hadoop.io.*;
>> -import org.apache.hadoop.mapred.*;
>> +import org.apache.hadoop.io.MapFile;
>> +import org.apache.hadoop.io.SequenceFile;
>> +import org.apache.hadoop.io.Text;
>> +import org.apache.hadoop.io.UTF8;
>> +import org.apache.hadoop.io.Writable;
>> +import org.apache.hadoop.io.WritableComparable;
>> +import org.apache.hadoop.mapred.FileSplit;
>> +import org.apache.hadoop.mapred.InputSplit;
>> +import org.apache.hadoop.mapred.JobClient;
>> +import org.apache.hadoop.mapred.JobConf;
>> +import org.apache.hadoop.mapred.Mapper;
>> +import org.apache.hadoop.mapred.OutputCollector;
>> +import org.apache.hadoop.mapred.OutputFormatBase;
>> +import org.apache.hadoop.mapred.RecordReader;
>> +import org.apache.hadoop.mapred.RecordWriter;
>> +import org.apache.hadoop.mapred.Reducer;
>> +import org.apache.hadoop.mapred.Reporter;
>> +import org.apache.hadoop.mapred.SequenceFileInputFormat;
>> +import org.apache.hadoop.mapred.SequenceFileRecordReader;
>>  import org.apache.hadoop.util.Progressable;
>>  import org.apache.nutch.crawl.CrawlDatum;
>>  import org.apache.nutch.crawl.Generator;
>> @@ -39,6 +59,7 @@
>>  import org.apache.nutch.parse.ParseText;
>>  import org.apache.nutch.protocol.Content;
>>  import org.apache.nutch.util.NutchConfiguration;
>> +import org.apache.nutch.util.NutchJob;
>>  
>>  /**
>>   * This tool takes several segments and merges their data together. Only the
>> @@ -482,7 +503,7 @@
>>      if (LOG.isInfoEnabled()) {
>>        LOG.info("Merging " + segs.length + " segments to " + out + "/" +
>> segmentName);
>>      }
>> -    JobConf job = new JobConf(getConf());
>> +    JobConf job = new NutchJob(getConf());
>>      job.setJobName("mergesegs " + out + "/" + segmentName);
>>      job.setBoolean("segment.merger.filter", filter);
>>      job.setLong("segment.merger.slice", slice);
>>
>> Modified: 
>> lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>> URL: 
>> http://svn.apache.org/viewvc/lucene/nutch/trunk/src/java/org/apache/nutch/segm
>> ent/SegmentReader.java?view=diff&rev=524932&r1=524931&r2=524932
>> ==============================================================================
>> --- lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>> (original)
>> +++ lucene/nutch/trunk/src/java/org/apache/nutch/segment/SegmentReader.java
>> Mon Apr  2 14:40:10 2007
>> @@ -17,18 +17,48 @@
>>  
>>  package org.apache.nutch.segment;
>>  
>> -import java.io.*;
>> +import java.io.BufferedReader;
>> +import java.io.BufferedWriter;
>> +import java.io.IOException;
>> +import java.io.InputStreamReader;
>> +import java.io.OutputStreamWriter;
>> +import java.io.PrintStream;
>> +import java.io.PrintWriter;
>> +import java.io.Writer;
>>  import java.text.SimpleDateFormat;
>> -import java.util.*;
>> +import java.util.ArrayList;
>> +import java.util.Arrays;
>> +import java.util.Date;
>> +import java.util.HashMap;
>> +import java.util.Iterator;
>> +import java.util.List;
>> +import java.util.Map;
>>  
>>  import org.apache.commons.logging.Log;
>>  import org.apache.commons.logging.LogFactory;
>> -
>>  import org.apache.hadoop.conf.Configuration;
>>  import org.apache.hadoop.conf.Configured;
>> -import org.apache.hadoop.fs.*;
>> -import org.apache.hadoop.io.*;
>> -import org.apache.hadoop.mapred.*;
>> +import org.apache.hadoop.fs.FileSystem;
>> +import org.apache.hadoop.fs.Path;
>> +import org.apache.hadoop.fs.PathFilter;
>> +import org.apache.hadoop.io.MapFile;
>> +import org.apache.hadoop.io.ObjectWritable;
>> +import org.apache.hadoop.io.SequenceFile;
>> +import org.apache.hadoop.io.Text;
>> +import org.apache.hadoop.io.UTF8;
>> +import org.apache.hadoop.io.Writable;
>> +import org.apache.hadoop.io.WritableComparable;
>> +import org.apache.hadoop.mapred.JobClient;
>> +import org.apache.hadoop.mapred.JobConf;
>> +import org.apache.hadoop.mapred.MapFileOutputFormat;
>> +import org.apache.hadoop.mapred.MapReduceBase;
>> +import org.apache.hadoop.mapred.Mapper;
>> +import org.apache.hadoop.mapred.OutputCollector;
>> +import org.apache.hadoop.mapred.RecordWriter;
>> +import org.apache.hadoop.mapred.Reducer;
>> +import org.apache.hadoop.mapred.Reporter;
>> +import org.apache.hadoop.mapred.SequenceFileInputFormat;
>> +import org.apache.hadoop.mapred.SequenceFileOutputFormat;
>>  import org.apache.hadoop.util.Progressable;
>>  import org.apache.nutch.crawl.CrawlDatum;
>>  import org.apache.nutch.parse.ParseData;
>> @@ -36,6 +66,7 @@
>>  import org.apache.nutch.protocol.Content;
>>  import org.apache.nutch.util.LogUtil;
>>  import org.apache.nutch.util.NutchConfiguration;
>> +import org.apache.nutch.util.NutchJob;
>>  
>>  /** Dump the content of a segment. */
>>  public class SegmentReader extends Configured implements Reducer {
>> @@ -120,7 +151,7 @@
>>    }
>>  
>>    private JobConf createJobConf() {
>> -    JobConf job = new JobConf(getConf());
>> +    JobConf job = new NutchJob(getConf());
>>      job.setBoolean("segment.reader.co", this.co);
>>      job.setBoolean("segment.reader.fe", this.fe);
>>      job.setBoolean("segment.reader.ge", this.ge);
>>
>>
> 
> 

-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
Nutch-developers mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-developers

Reply via email to