No, not using any script of that sort. Sure.

Thanks.
On 12 Jul 2016 22:08, "Tom Crayford" <tcrayf...@heroku.com> wrote:

> Ah, that's around leader rebalancing. Do you have any scripts that run
> kafka-assign-partitions or similar?
>
> I will recheck but this doesn't sound like a thing that auto rebalance
> would impact
>
> On Tuesday, 12 July 2016, Gokul <gokulakanna...@gmail.com> wrote:
>
>> Thanks. Auto rebalance is set to true, so rebalancing may be happening at
>> that time. Is there any issue tracker that I can refer to?
>> On 12 Jul 2016 21:48, "Tom Crayford" <tcrayf...@heroku.com> wrote:
>>
>>> Hi,
>>>
>>> Were you rebalancing that topic or partition at that time? If there are
>>> rebalancing bugs this might point at that.
>>>
>>> Thanks
>>>
>>> Tom
>>>
>>> On Tue, Jul 12, 2016 at 6:47 AM, Gokul <slu...@gmail.com> wrote:
>>>
>>> > We had an issue last week when kafka cluster reported under replicated
>>> > partitions for quite a while but there were no brokers down. All the
>>> > brokers were reporting unknownException on Broker 1. When checked
>>> broker 1
>>> > logs, it just reported below errors(NotAssignedReplicaException)
>>> > continuously. Issue got resolved after bouncing broker 1. Think this
>>> > exception comes when Controller issues StopReplicaRequest to broker 1
>>> and
>>> > it is in the process of leader election. But what is spooky is that
>>> this
>>> > exception was reported more than 20 minutes by the broker and it was
>>> > chocking entire ingestion(totally 10 brokers). Unfortunately I don't
>>> have
>>> > the controller logs to debug. Any pointers here? We are using 0.8.2.1
>>> >
>>> > ERROR [2016-07-07 11:45:09,248] [kafka-request-handler-1][]
>>> > kafka.server.KafkaApis - [KafkaApi-1] error when handling request Name:
>>> > FetchRequest; Version: 0; CorrelationId: 1890972; ClientId:
>>> > ReplicaFetcherThread-2-1; ReplicaId: 2; MaxWait: 500 ms; MinBytes: 1
>>> bytes;
>>> > RequestInfo: <value>
>>> > kafka.common.NotAssignedReplicaException: Leader 1 failed to record
>>> > follower 2's position 20240372 since the replica is not recognized to
>>> be
>>> > one of the assigned replicas 1,5,6 for partition [<topic>,1]
>>> >         at
>>> >
>>> >
>>> kafka.server.ReplicaManager.updateReplicaLEOAndPartitionHW(ReplicaManager.scala:574)
>>> >         at
>>> >
>>> >
>>> kafka.server.KafkaApis$$anonfun$recordFollowerLogEndOffsets$2.apply(KafkaApis.scala:388)
>>> >         at
>>> >
>>> >
>>> kafka.server.KafkaApis$$anonfun$recordFollowerLogEndOffsets$2.apply(KafkaApis.scala:386)
>>> >         at
>>> >
>>> >
>>> scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
>>> >         at
>>> >
>>> >
>>> scala.collection.MapLike$MappedValues$$anonfun$foreach$3.apply(MapLike.scala:245)
>>> >         at
>>> >
>>> >
>>> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:778)
>>> >         at
>>> > scala.collection.immutable.HashMap$HashMap1.foreach(HashMap.scala:221)
>>> >         at
>>> >
>>> scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:428)
>>> >         at
>>> >
>>> scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:428)
>>> >         at
>>> >
>>> >
>>> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:777)
>>> >         at
>>> scala.collection.MapLike$MappedValues.foreach(MapLike.scala:245)
>>> >         at
>>> > kafka.server.KafkaApis.recordFollowerLogEndOffsets(KafkaApis.scala:386)
>>> >         at
>>> kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:351)
>>> >         at kafka.server.KafkaApis.handle(KafkaApis.scala:60)
>>> >         at
>>> > kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:59)
>>> >         at java.lang.Thread.run(Thread.java:745)
>>> >
>>> > --
>>> > Thanks and Regards,
>>> > Gokul
>>> >
>>>
>>

Reply via email to