Thanks Michael! I just committed your fix. It will be released with 4.7.1

On Fri, Mar 21, 2014 at 8:30 PM, Michael Sokolov
<msoko...@safaribooksonline.com> wrote:
> I just managed to track this down -- as you said the disconnect was a red
> herring.
>
> Ultimately the problem was caused by a custom analysis component we wrote
> that was raising an IOException -- it was missing some configuration files
> it relies on.
>
> What might be interesting for solr devs to have a look at is that exception
> was completely swallowed by JavabinCodec, making it very difficult to track
> down the problem.  Furthermore -- if the /add request was routed directly to
> the shard where the document was destined to end up, then the IOException
> raised by the analysis component (a char filter) showed up in the Solr HTTP
> response (probably because my client used XML format in one test -- javabin
> is used internally in SolrCloud).  But if the request was routed to a
> different shard, then the only exception that showed up anywhere (in the
> logs, in the HTTP response) was kind of irrelevant.
>
> I think this could be fixed pretty easily; see SOLR-5985 for my suggestion.
>
> -Mike
>
>
>
> On 03/21/2014 10:20 AM, Greg Walters wrote:
>>
>> Broken pipe errors are generally caused by unexpected disconnections and
>> are some times hard to track down. Given the stack traces you've provided
>> it's hard to point to any one thing and I suspect the relevant information
>> was snipped out in the "long dump of document fields". You might grab the
>> entire error from the client you're uploading documents with, the server
>> you're connected to and any other nodes that have an error at the same time
>> and put it on pastebin or the like.
>>
>> Thanks,
>> Greg
>>
>> On Mar 20, 2014, at 3:36 PM, Michael Sokolov
>> <msoko...@safaribooksonline.com> wrote:
>>
>>> I'm getting a similar exception when writing documents (on the client
>>> side).  I can write one document fine, but the second (which is being routed
>>> to a different shard) generates the error.  It happens every time -
>>> definitely not a resource issue or timing problem since this database is
>>> completely empty -- I'm just getting started and running some tests, so
>>> there must be some kind of setup problem.  But it's difficult to diagnose
>>> (for me, anyway)!  I'd appreciate any insight, hints, guesses, etc. since
>>> I'm stuck. Thanks!
>>>
>>> One node (the leader?) is reporting "Internal Server Error" in its log,
>>> and another node (presumably the shard where the document is being directed)
>>> bombs out like this:
>>>
>>> ERROR - 2014-03-20 15:56:53.022; org.apache.solr.common.SolrException;
>>> null:org.apache.solr.common.SolrException: ERROR adding document
>>> SolrInputDocument(
>>>
>>> ... long dump of document fields
>>>
>>> )
>>>     at
>>> org.apache.solr.handler.loader.JavabinLoader$1.update(JavabinLoader.java:99)
>>>     at
>>> org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readOuterMostDocIterator(JavaBinUpdateRequestCodec.java:166)
>>>     at
>>> org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readIterator(JavaBinUpdateRequestCodec.java:136)
>>>     at
>>> org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:225)
>>>     at
>>> org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec$1.readNamedList(JavaBinUpdateRequestCodec.java:121)
>>>     at
>>> org.apache.solr.common.util.JavaBinCodec.readVal(JavaBinCodec.java:190)
>>>     at
>>> org.apache.solr.common.util.JavaBinCodec.unmarshal(JavaBinCodec.java:116)
>>>     at
>>> org.apache.solr.client.solrj.request.JavaBinUpdateRequestCodec.unmarshal(JavaBinUpdateRequestCodec.java:173)
>>>     at
>>> org.apache.solr.handler.loader.JavabinLoader.parseAndLoadDocs(JavabinLoader.java:106)
>>>     at
>>> org.apache.solr.handler.loader.JavabinLoader.load(JavabinLoader.java:58)
>>>     at
>>> org.apache.solr.handler.UpdateRequestHandler$1.load(UpdateRequestHandler.java:92)
>>>     at
>>> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:74)
>>>     at
>>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
>>>     at org.apache.solr.core.SolrCore.execute(SolrCore.java:1859)
>>>     at
>>> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:721)
>>> ...
>>> Caused by: java.net.SocketException: Broken pipe
>>>         at java.net.SocketOutputStream.socketWrite0(Native Method)
>>>         at
>>> java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:109)
>>>         at java.net.SocketOutputStream.write(SocketOutputStream.java:153)
>>>         at
>>> org.apache.coyote.http11.InternalOutputBuffer.realWriteBytes(InternalOutputBuffer.java:215)
>>>         at
>>> org.apache.tomcat.util.buf.ByteChunk.flushBuffer(ByteChunk.java:480)
>>>         at
>>> org.apache.tomcat.util.buf.ByteChunk.append(ByteChunk.java:366)
>>>         at
>>> org.apache.coyote.http11.InternalOutputBuffer$OutputStreamOutputBuffer.doWrite(InternalOutputBuffer.java:240)
>>>         at
>>> org.apache.coyote.http11.filters.ChunkedOutputFilter.doWrite(ChunkedOutputFilter.java:119)
>>>         at
>>> org.apache.coyote.http11.AbstractOutputBuffer.doWrite(AbstractOutputBuffer.java:192)
>>>         at org.apache.coyote.Response.doWrite(Response.java:520)
>>>         at
>>> org.apache.catalina.connector.OutputBuffer.realWriteBytes(OutputBuffer.java:408)
>>>         ... 37 more
>>>
>>> This is with Solr 4.6.1, Tomcat 7.  Here's my clusterstate.json. Updates
>>> are being sent to the test1x3 collection
>>>
>>>
>>> {
>>>   "test3x1":{
>>>     "shards":{
>>>       "shard1":{
>>>         "range":"80000000-d554ffff",
>>>         "state":"active",
>>>         "replicas":{"core_node1":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.37:8080/solr";,
>>>             "core":"test3x1_shard1_replica1",
>>>             "node_name":"10.4.24.37:8080_solr",
>>>             "leader":"true"}}},
>>>       "shard2":{
>>>         "range":"d5550000-2aa9ffff",
>>>         "state":"active",
>>>         "replicas":{"core_node3":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.39:8080/solr";,
>>>             "core":"test3x1_shard2_replica1",
>>>             "node_name":"10.4.24.39:8080_solr",
>>>             "leader":"true"}}},
>>>       "shard3":{
>>>         "range":"2aaa0000-7fffffff",
>>>         "state":"active",
>>>         "replicas":{"core_node2":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.38:8080/solr";,
>>>             "core":"test3x1_shard3_replica1",
>>>             "node_name":"10.4.24.38:8080_solr",
>>>             "leader":"true"}}}},
>>>     "maxShardsPerNode":"1",
>>>     "router":{"name":"compositeId"},
>>>     "replicationFactor":"1"},
>>>   "test1x3":{
>>>     "shards":{"shard1":{
>>>         "range":"80000000-7fffffff",
>>>         "state":"active",
>>>         "replicas":{
>>>           "core_node1":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.39:8080/solr";,
>>>             "core":"test1x3_shard1_replica2",
>>>             "node_name":"10.4.24.39:8080_solr",
>>>             "leader":"true"},
>>>           "core_node2":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.38:8080/solr";,
>>>             "core":"test1x3_shard1_replica1",
>>>             "node_name":"10.4.24.38:8080_solr"},
>>>           "core_node3":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.37:8080/solr";,
>>>             "core":"test1x3_shard1_replica3",
>>>             "node_name":"10.4.24.37:8080_solr"}}}},
>>>     "maxShardsPerNode":"1",
>>>     "router":{"name":"compositeId"},
>>>     "replicationFactor":"3"},
>>>   "test2x2":{
>>>     "shards":{
>>>       "shard1":{
>>>         "range":"80000000-ffffffff",
>>>         "state":"active",
>>>         "replicas":{
>>>           "core_node1":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.39:8080/solr";,
>>>             "core":"test2x2_shard1_replica1",
>>>             "node_name":"10.4.24.39:8080_solr"},
>>>           "core_node4":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.38:8080/solr";,
>>>             "core":"test2x2_shard1_replica2",
>>>             "node_name":"10.4.24.38:8080_solr",
>>>             "leader":"true"}}},
>>>       "shard2":{
>>>         "range":"0-7fffffff",
>>>         "state":"active",
>>>         "replicas":{
>>>           "core_node2":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.37:8080/solr";,
>>>             "core":"test2x2_shard2_replica1",
>>>             "node_name":"10.4.24.37:8080_solr",
>>>             "leader":"true"},
>>>           "core_node3":{
>>>             "state":"active",
>>>             "base_url":"http://10.4.24.39:8080/solr";,
>>>             "core":"test2x2_shard2_replica2",
>>>             "node_name":"10.4.24.39:8080_solr"}}}},
>>>     "maxShardsPerNode":"2",
>>>     "router":{"name":"compositeId"},
>>>     "replicationFactor":"2"}}
>>>
>>>
>>>
>>> On 03/20/2014 09:44 AM, Greg Walters wrote:
>>>>
>>>> Sathya,
>>>>
>>>> I assume you're using Solr Cloud. Please provide your clusterstate.json
>>>> while you're seeing this issue and check your logs for any exceptions. With
>>>> no information from you it's hard to troubleshoot any issues!
>>>>
>>>> Thanks,
>>>> Greg
>>>>
>>>> On Mar 20, 2014, at 12:44 AM, Sathya <sathia.blacks...@gmail.com> wrote:
>>>>
>>>>> Hi Friends,
>>>>>
>>>>> I am new to Solr. I have 5 solr node in 5 different machine. When i
>>>>> index
>>>>> the data, sometimes "*No live SolrServers available to handle this
>>>>> request*"
>>>>> exception occur in 1 or 2 machines.
>>>>>
>>>>> I dont know why its happen and how to solve this. Kindly help me to
>>>>> solve
>>>>> this issue.
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> View this message in context:
>>>>> http://lucene.472066.n3.nabble.com/Solr4-7-No-live-SolrServers-available-to-handle-this-request-tp4125679.html
>>>>> Sent from the Solr - User mailing list archive at Nabble.com.
>
>



-- 
Regards,
Shalin Shekhar Mangar.

Reply via email to