Hi,
I just upgraded from trunk version 28.12.2008 to trunk version 04.02.2009.
Now, I'm trying to read my old crawldb's e.g. by using the command "bin/nutch
readdb <crawldb> -stats" , but I always get the following error:
2009-02-10 15:41:05,541 DEBUG mapred.MapTask - Writing local split to
/tmp/CRAWLNAME.default.xyz/mapred/local/localRunner/split.dta
2009-02-10 15:41:05,588 DEBUG mapred.TaskRunner - attempt_local_0001_m_000000_0
Progress/ping thread started
2009-02-10 15:41:05,588 INFO mapred.MapTask - numReduceTasks: 1
2009-02-10 15:41:05,588 INFO mapred.MapTask - io.sort.mb = 100
2009-02-10 15:41:05,698 INFO mapred.MapTask - data buffer = 79691776/99614720
2009-02-10 15:41:05,698 INFO mapred.MapTask - record buffer = 262144/327680
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Creating group
org.apache.hadoop.mapred.Task$Counter with bundle
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding MAP_OUTPUT_BYTES
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding MAP_OUTPUT_RECORDS
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding COMBINE_INPUT_RECORDS
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding COMBINE_OUTPUT_RECORDS
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding MAP_INPUT_RECORDS
2009-02-10 15:41:05,713 DEBUG mapred.Counters - Adding MAP_INPUT_BYTES
2009-02-10 15:41:05,729 WARN mapred.LocalJobRunner - job_local_0001
java.lang.RuntimeException: java.lang.NullPointerException
at
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:81)
at
org.apache.hadoop.io.MapWritable.readFields(MapWritable.java:164)
at
org.apache.nutch.crawl.CrawlDatum.readFields(CrawlDatum.java:262)
at
org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
at
org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
at
org.apache.hadoop.io.SequenceFile$Reader.deserializeValue(SequenceFile.java:1817)
at
org.apache.hadoop.io.SequenceFile$Reader.getCurrentValue(SequenceFile.java:1790)
at
org.apache.hadoop.mapred.SequenceFileRecordReader.getCurrentValue(SequenceFileRecordReader.java:103)
at
org.apache.hadoop.mapred.SequenceFileRecordReader.next(SequenceFileRecordReader.java:78)
at
org.apache.hadoop.mapred.MapTask$TrackedRecordReader.moveToNext(MapTask.java:186)
at
org.apache.hadoop.mapred.MapTask$TrackedRecordReader.next(MapTask.java:170)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:48)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:138)
Caused by: java.lang.NullPointerException
at
java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:768)
at
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:73)
... 13 more
With the older version oft he trunk I can read the crawldb without difficulty.
Are the old files not readable with the new trunk version since the upgrade to
lucene 2.4?
Is there anything I can do to re-use my old data with the new version?
Kind regards,
Martina