Hi Brian,
I've never seen this before.
I found this however

http://web.archiveorange.com/archive/v/L9Ul807Yu77D5QW7PGPn

I know posting links to resolve problems is not ideal... bit as I said I've
never seen it before. Interesting thought that this happens intermittently
same as in the issue closed off...

On Tuesday, August 13, 2013, brian4 <[email protected]> wrote:
> I am running nutch 2.2.1 (not distributed) with hbase and solr 3.5.  When
I
> run 10-15 crawls in a row, all on one machine, it seems one crawl will
> randomly fail at index time with the following log messages, but the
> subsequent crawls will work and index fine:
>
> 013-08-13 12:58:57,430 INFO  collection.CollectionManager - file has23
> elements
> 2013-08-13 12:58:58,914 INFO  solr.SolrWriter - Adding 158 documents
> 2013-08-13 12:58:59,034 INFO  httpclient.HttpMethodDirector - I/O
exception
> (java.net.SocketException) caught when processing request: Connection
reset
> 2013-08-13 12:58:59,035 INFO  httpclient.HttpMethodDirector - Retrying
> request
> 2013-08-13 12:58:59,037 INFO  solr.SolrWriter - Adding 158 documents
> 2013-08-13 12:58:59,076 INFO  httpclient.HttpMethodDirector - I/O
exception
> (java.net.SocketException) caught when processing request: Connection
reset
> 2013-08-13 12:58:59,076 INFO  httpclient.HttpMethodDirector - Retrying
> request
> 2013-08-13 12:58:59,077 WARN  mapred.FileOutputCommitter - Output path is
> null in cleanup
> 2013-08-13 12:58:59,078 WARN  mapred.LocalJobRunner -
> job_local899249969_0001
> java.lang.Exception: java.io.IOException:
> org.apache.solr.client.solrj.SolrServerException:
> org.apache.commons.httpclient.ProtocolException: Unbuffered entity
enclosing
> request can not be repeated.
>         at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:354)
> Caused by: java.io.IOException:
> org.apache.solr.client.solrj.SolrServerException:
> org.apache.commons.httpclient.ProtocolException: Unbuffered entity
enclosing
> request can not be repeated.
>         at
org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:95)
>         at
>
org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:53)
>         at
>
org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.close(MapTask.java:650)
>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:767)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:364)
>         at
>
org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:223)
>         at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>         at
java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:166)
>         at
>
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
>         at
>
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
>         at java.lang.Thread.run(Thread.java:722)
> Caused by: org.apache.solr.client.solrj.SolrServerException:
> org.apache.commons.httpclient.ProtocolException: Unbuffered entity
enclosing
> request can not be repeated.
>         at
>
org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:475)
>         at
>
org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:244)
>         at
>
org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:105)
>         at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:49)
>         at
org.apache.nutch.indexer.solr.SolrWriter.close(SolrWriter.java:91)
>         ... 11 more
> Caused by: org.apache.commons.httpclient.ProtocolException: Unbuffered
> entity enclosing request can not be repeated.
>         at
>
org.apache.commons.httpclient.methods.EntityEnclosingMethod.writeRequestBody(EntityEnclosingMethod.java:487)
>         at
>
org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2114)
>         at
>
org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096)
>         at
>
org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398)
>         at
>
org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171)
>         at
>
org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397)
>         at
>
org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323)
>         at
>
org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:422)
>         ... 15 more
> 2013-08-13 12:58:59,640 ERROR solr.SolrIndexerJob - SolrIndexerJob:
> java.lang.RuntimeException: job failed: name=[events_crawl]solr-index,
> jobid=job_local899249969_0001
>         at
org.apache.nutch.util.NutchJob.waitForCompletion(NutchJob.java:54)
>         at
org.apache.nutch.indexer.solr.SolrIndexerJob.run(SolrIndexerJob.java:46)
>         at
>
org.apache.nutch.indexer.solr.SolrIndexerJob.indexSolr(SolrIndexerJob.java:54)
>         at
org.apache.nutch.indexer.solr.SolrIndexerJob.run(SolrIndexerJob.java:76)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>         at
> org.apache.nutch.indexer.solr.SolrIndexerJob.main(SolrIndexerJob.java:85)
>
> Any ideas why?  Even if for some reason the server resets the connection
> randomly, it should be able to reconnect and continue adding documents
> right?  Maybe I need to modify the solr indexer code to reconnect?
>
> I saw the following similar issue in 1.x, but it was dropped as not
> reproducible.  However I can reproduce this fairly consistently.
>
> https://issues.apache.org/jira/browse/NUTCH-1348
>
>
>
> --
> View this message in context:
http://lucene.472066.n3.nabble.com/SolrIndexerJob-connection-reset-job-failed-tp4084373.html
> Sent from the Nutch - User mailing list archive at Nabble.com.
>

-- 
*Lewis*

Reply via email to