Well, it does not look like I can reproduce it anymore. While I was able to 
consistently crash it
yesterday, today kafka broker runs without any problems.

--
Dmitri Priimak

On 11/29/2012 02:10 PM, Dmitri Priimak wrote:
> On 11/29/2012 02:06 PM, Neha Narkhede wrote:
>> That just meant that we couldn't reproduce it during our testing, but
>> we occasionally do see it in production, which is much harder to
>> reproduce.
>> It will be great if you can reproduce the issue and attach the test case.
> I see. I will try to isolate this problem. I am hoping that I can locate 
> exact sequence of messages
> that trigger this problem, but at the moment I cannot easily do that.
>
> --
> Dmitri Priimak
>
>> Thanks,
>> Neha
>>
>> On Thu, Nov 29, 2012 at 2:03 PM, Dmitri Priimak
>> <prii...@highwire.stanford.edu> wrote:
>>> On 11/29/2012 01:50 PM, Neha Narkhede wrote:
>>>> Dmitri,
>>>>
>>>> You said that the problem goes away once you upgrade to 0.7.1. Do you
>>>> mean that the message corruption doesn't happen or that when it
>>>> happens, the broker doesn't shut down ?
>>> I do not see corrupt message at all.
>>>
>>>
>>>> I'm asking since we occasionally do see the corruption on 0.7.1 but
>>>> don't have a good way to reproduce that. 0.7.1 fixes the shut down
>>>> issue so the broker ignores corrupted requests
>>>> instead of shutting down - https://issues.apache.org/jira/browse/KAFKA-261
>>>>
>>>> Do you have a way to reproduce this issue ?
>>> Sort of.
>>>
>>> I see this bug report https://issues.apache.org/jira/browse/KAFKA-273 that 
>>> says "Unresolved"
>>> and it shows the same stacktrace as the one I reported. However that 
>>> comments at the bottom
>>> sounds like this issue might have been resolved. Do you know if that is the 
>>> case?
>>>
>>> --
>>> Dmitri Priimak
>>>
>>>> It will really help in
>>>> troubleshooting the root cause.
>>>>
>>>> Thanks,
>>>> Neha
>>>>
>>>> On Thu, Nov 29, 2012 at 8:37 AM, Dmitri Priimak
>>>> <prii...@highwire.stanford.edu> wrote:
>>>>> On 12-11-28 09:54 PM, Jun Rao wrote:
>>>>>> Dmitri,
>>>>>>
>>>>>> Could you reproduce this easily? Are you using a load balancer? Earlier,
>>>>>> another user had the same issue and eventually figured out that the
>>>>>> problem
>>>>>> is in the network router.
>>>>> I do not believe I have any loadbalancer anywhere in the picture and by 
>>>>> the
>>>>> way problem does go
>>>>> away if I upgrade broker to version 0.7.1 or 0.7.2
>>>>>
>>>>> --
>>>>> Dmitri Priimak
>>>>>
>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> Jun
>>>>>>
>>>>>> On Wed, Nov 28, 2012 at 11:34 AM, Dmitri Priimak <
>>>>>> prii...@highwire.stanford.edu> wrote:
>>>>>>
>>>>>>> Hi.
>>>>>>>
>>>>>>> In the kafka broker (version 0.7.0) log I see occasionally following
>>>>>>> error
>>>>>>> message
>>>>>>>
>>>>>>>   FATAL Halting due to unrecoverable I/O error while handling producer
>>>>>>> request: Unexpected end of
>>>>>>> ZLIB input stream (kafka.server.KafkaRequestHandlers)
>>>>>>> java.io.EOFException: Unexpected end of ZLIB input stream
>>>>>>>          at
>>>>>>> java.util.zip.InflaterInputStream.fill(InflaterInputStream.java:223)
>>>>>>>          at
>>>>>>> java.util.zip.InflaterInputStream.read(InflaterInputStream.java:141)
>>>>>>>          at java.util.zip.GZIPInputStream.read(GZIPInputStream.java:90)
>>>>>>>          at java.io.FilterInputStream.read(FilterInputStream.java:90)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.CompressionUtils$$anonfun$decompress$4.apply$mcI$sp(CompressionUtils.scala:123)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.CompressionUtils$$anonfun$decompress$4.apply(CompressionUtils.scala:123)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.CompressionUtils$$anonfun$decompress$4.apply(CompressionUtils.scala:123)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.immutable.Stream$$anonfun$continually$1.apply(Stream.scala:598)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.immutable.Stream$$anonfun$continually$1.apply(Stream.scala:598)
>>>>>>>          at 
>>>>>>> scala.collection.immutable.Stream$Cons.tail(Stream.scala:555)
>>>>>>>          at 
>>>>>>> scala.collection.immutable.Stream$Cons.tail(Stream.scala:549)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.immutable.Stream$$anonfun$takeWhile$1.apply(Stream.scala:394)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.immutable.Stream$$anonfun$takeWhile$1.apply(Stream.scala:394)
>>>>>>>          at 
>>>>>>> scala.collection.immutable.Stream$Cons.tail(Stream.scala:555)
>>>>>>>          at 
>>>>>>> scala.collection.immutable.Stream$Cons.tail(Stream.scala:549)
>>>>>>>          at scala.collection.immutable.Stream.foreach(Stream.scala:255)
>>>>>>>          at
>>>>>>> kafka.message.CompressionUtils$.decompress(CompressionUtils.scala:123)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.ByteBufferMessageSet$$anon$1.makeNextOuter(ByteBufferMessageSet.scala:124)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.ByteBufferMessageSet$$anon$1.makeNext(ByteBufferMessageSet.scala:138)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.message.ByteBufferMessageSet$$anon$1.makeNext(ByteBufferMessageSet.scala:82)
>>>>>>>          at
>>>>>>> kafka.utils.IteratorTemplate.maybeComputeNext(IteratorTemplate.scala:59)
>>>>>>>          at
>>>>>>> kafka.utils.IteratorTemplate.hasNext(IteratorTemplate.scala:51)
>>>>>>>          at scala.collection.Iterator$class.foreach(Iterator.scala:631)
>>>>>>>          at
>>>>>>> kafka.utils.IteratorTemplate.foreach(IteratorTemplate.scala:30)
>>>>>>>          at
>>>>>>> scala.collection.IterableLike$class.foreach(IterableLike.scala:79)
>>>>>>>          at kafka.message.MessageSet.foreach(MessageSet.scala:87)
>>>>>>>          at kafka.log.Log.append(Log.scala:202)
>>>>>>>          at
>>>>>>>
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers.kafka$server$KafkaRequestHandlers$$handleProducerRequest(KafkaRequestHandlers.scala:75)
>>>>>>>          at
>>>>>>>
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers$$anonfun$handleMultiProducerRequest$1.apply(KafkaRequestHandlers.scala:68)
>>>>>>>          at
>>>>>>>
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers$$anonfun$handleMultiProducerRequest$1.apply(KafkaRequestHandlers.scala:68)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
>>>>>>>          at
>>>>>>>
>>>>>>> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34)
>>>>>>>          at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:34)
>>>>>>>          at
>>>>>>> scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
>>>>>>>          at scala.collection.mutable.ArrayOps.map(ArrayOps.scala:34)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers.handleMultiProducerRequest(KafkaRequestHandlers.scala:68)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$4.apply(KafkaRequestHandlers.scala:46)
>>>>>>>          at
>>>>>>>
>>>>>>> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$4.apply(KafkaRequestHandlers.scala:46)
>>>>>>>          at kafka.network.Processor.handle(SocketServer.scala:289)
>>>>>>>          at kafka.network.Processor.read(SocketServer.scala:312)
>>>>>>>          at kafka.network.Processor.run(SocketServer.scala:207)
>>>>>>>          at java.lang.Thread.run(Thread.java:662)
>>>>>>>
>>>>>>> At which point broker actually dies. Shouldn't it keep working even even
>>>>>>> if there is some such error?
>>>>>>> Also, does anyone else saw this error? And is it fixed in the newer
>>>>>>> versions?
>>>>>>>
>>>>>>> --
>>>>>>> Dmitri Priimak
>>>>>>>
>>>>>>>

Reply via email to