Your problem is not the same judging from the stack trace on the Solr list. 
Your Solr runs OOM, not Nutch.

On Thursday 27 October 2011 14:20:10 Fred Zimmerman wrote:
> I'm having the exact same problem. I am trying to isolate whether it is a
> Solr problem or a Nutch+Solr problem.
> 
> On Wed, Oct 26, 2011 at 11:54 PM, <[email protected]> wrote:
> > Hi,
> > 
> > I am working with a Nutch 1.4 snapshot and having a very strange problem
> > that makes the system run out of memory when indexing into Solr. This
> > does not look like a trivial lack of memory problem that can be solved
> > by giving more memory to the JVM. I've increased the max memory size
> > from 2Gb to 3Gb, then to 6Gb, but this did not make any difference.
> > 
> > A log extract is included below.
> > 
> > Would anyone have any idea of how to fix this problem?
> > 
> > Thanks,
> > 
> > Arkadi
> > 
> > 
> > 2011-10-27 07:08:22,162 INFO  solr.SolrWriter - Adding 1000 documents
> > 2011-10-27 07:08:42,248 INFO  solr.SolrWriter - Adding 1000 documents
> > 2011-10-27 07:13:54,110 WARN  mapred.LocalJobRunner - job_local_0254
> > java.lang.OutOfMemoryError: Java heap space
> > 
> >       at java.util.Arrays.copyOfRange(Arrays.java:3209)
> >       at java.lang.String.<init>(String.java:215)
> >       at java.nio.HeapCharBuffer.toString(HeapCharBuffer.java:542)
> >       at java.nio.CharBuffer.toString(CharBuffer.java:1157)
> >       at org.apache.hadoop.io.Text.decode(Text.java:350)
> >       at org.apache.hadoop.io.Text.decode(Text.java:322)
> >       at org.apache.hadoop.io.Text.readString(Text.java:403)
> >       at org.apache.nutch.parse.ParseText.readFields(ParseText.java:50)
> >       at
> > 
> > org.apache.nutch.util.GenericWritableConfigurable.readFields(GenericWrita
> > bleConfigurable.java:54)
> > 
> >       at
> > 
> > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserialize
> > r.deserialize(WritableSerialization.java:67)
> > 
> >       at
> > 
> > org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserialize
> > r.deserialize(WritableSerialization.java:40)
> > 
> >       at
> > 
> > org.apache.hadoop.mapred.Task$ValuesIterator.readNextValue(Task.java:991)
> > 
> >       at org.apache.hadoop.mapred.Task$ValuesIterator.next(Task.java:931)
> >       at
> > 
> > org.apache.hadoop.mapred.ReduceTask$ReduceValuesIterator.moveToNext(Reduc
> > eTask.java:241)
> > 
> >       at
> > 
> > org.apache.hadoop.mapred.ReduceTask$ReduceValuesIterator.next(ReduceTask.
> > java:237)
> > 
> >       at
> > 
> > org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:81
> > )
> > 
> >       at
> > 
> > org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:50
> > )
> > 
> >       at
> > 
> > org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:463)
> > 
> >       at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411)
> >       at
> > 
> > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:216)
> > 2011-10-27 07:13:54,382 ERROR solr.SolrIndexer - java.io.IOException: Job
> > failed!

-- 
Markus Jelsma - CTO - Openindex
http://www.linkedin.com/in/markus17
050-8536620 / 06-50258350

Reply via email to