It could mean that the request offset no longer exists in the broker since
it's too old. If you look at the log segments in the broker, you can see
the valid offset range. The first valid offset is part of the name of the
oldest log segment.

Thanks,

Jun


On Wed, Apr 9, 2014 at 12:59 AM, darion <[email protected]> wrote:

> Hi exports
> Storm Topology reading data from kafka and kafka just a singal node
> Topology Spout is so slow and I found this excption in the kafka's logging
>
> [2014-04-09 14:58:53,729] ERROR error when processing request
> FetchRequest(topic:topic.nginx, part:0 offset:948810259194
> maxSize:1048576) (kafka.server.KafkaRequestHandlers)
> kafka.common.OffsetOutOfRangeException: offset 948810259194 is out of range
> at kafka.log.Log$.findRange(Log.scala:46)
> at kafka.log.Log.read(Log.scala:264)
> at
>
> kafka.server.KafkaRequestHandlers.kafka$server$KafkaRequestHandlers$$readMessageSet(KafkaRequestHandlers.scala:112)
> at
>
> kafka.server.KafkaRequestHandlers$$anonfun$2.apply(KafkaRequestHandlers.scala:101)
> at
>
> kafka.server.KafkaRequestHandlers$$anonfun$2.apply(KafkaRequestHandlers.scala:100)
> at
>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
> at
>
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34)
> at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:34)
> at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
> at scala.collection.mutable.ArrayOps.map(ArrayOps.scala:34)
> at
>
> kafka.server.KafkaRequestHandlers.handleMultiFetchRequest(KafkaRequestHandlers.scala:100)
> at
>
> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$3.apply(KafkaRequestHandlers.scala:40)
> at
>
> kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$3.apply(KafkaRequestHandlers.scala:40)
> at kafka.network.Processor.handle(SocketServer.scala:296)
> at kafka.network.Processor.read(SocketServer.scala:319)
> at kafka.network.Processor.run(SocketServer.scala:214)
> at java.lang.Thread.run(Thread.java:662)
>
> It is meaning kafka's hold data too large ? or any other something?
> It's a bug ?
>
> thx a lot
>

Reply via email to