Yes, perhaps one fix would be to overload the offset MIN_LONG/MAX_LONG to mean first/last.
-Jay On Thu, Oct 6, 2011 at 8:47 AM, Jun Rao <jun...@gmail.com> wrote: > Thai, > > When a consumer starts off for the first time, it doesn't know which offset > to start with. So it picks max_long, which will trigger the > WrongOffsetException. Upon receiving the exception, the consumer will reset > its offset to either the smallest or the largest offset depending on the > configuration. So, technically, this is not really an error and I agree > that > it's confusing to log this as an error. Could you open a jira for this? > > Thanks, > > Jun > > On Thu, Oct 6, 2011 at 4:08 AM, Bao Thai Ngo <baothai...@gmail.com> wrote: > > > Hi, > > > > I followed the quickstart to setup and run kafka on my Laptop CentOS 5.5 > > and > > got this error: > > > > [2011-10-06 17:21:20,349] INFO Starting log flusher every 1000 ms with > the > > following overrides Map() (kafka.log.LogManager) > > [2011-10-06 17:21:20,350] INFO Server started. (kafka.server.KafkaServer) > > [2011-10-06 17:22:28,366] INFO Closing socket connection to /10.40.9.171 > . > > (kafka.network.Processor) > > [2011-10-06 17:38:53,725] INFO Closing socket connection to /127.0.0.1. > > (kafka.network.Processor) > > [2011-10-06 17:39:01,445] INFO Created log for 'testxx'-0 > > (kafka.log.LogManager) > > [2011-10-06 17:39:01,448] INFO Begin registering broker topic > > /brokers/topics/testxx/0 with 1 partitions (kafka.server.KafkaZooKeeper) > > [2011-10-06 17:39:01,472] INFO End registering broker topic > > /brokers/topics/testxx/0 (kafka.server.KafkaZooKeeper) > > [2011-10-06 17:39:50,641] ERROR error when processing request > > FetchRequest(topic:testxx, part:0 offset:9223372036854775807 > > maxSize:307200) > > (kafka.server.KafkaRequestHandlers) > > kafka.common.OffsetOutOfRangeException: offset 9223372036854775807 is out > > of > > range > > at kafka.log.Log$.findRange(Log.scala:48) > > at kafka.log.Log.read(Log.scala:224) > > at > > > > > kafka.server.KafkaRequestHandlers.kafka$server$KafkaRequestHandlers$$readMessageSet(KafkaRequestHandlers.scala:116) > > at > > > > > kafka.server.KafkaRequestHandlers$$anonfun$2.apply(KafkaRequestHandlers.scala:106) > > at > > > > > kafka.server.KafkaRequestHandlers$$anonfun$2.apply(KafkaRequestHandlers.scala:105) > > at > > > > > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) > > at > > > > > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206) > > at > > > > > scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:34) > > at scala.collection.mutable.ArrayOps.foreach(ArrayOps.scala:34) > > at > > scala.collection.TraversableLike$class.map(TraversableLike.scala:206) > > at scala.collection.mutable.ArrayOps.map(ArrayOps.scala:34) > > at > > > > > kafka.server.KafkaRequestHandlers.handleMultiFetchRequest(KafkaRequestHandlers.scala:105) > > at > > > > > kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$3.apply(KafkaRequestHandlers.scala:45) > > at > > > > > kafka.server.KafkaRequestHandlers$$anonfun$handlerFor$3.apply(KafkaRequestHandlers.scala:45) > > at kafka.network.Processor.handle(SocketServer.scala:289) > > at kafka.network.Processor.read(SocketServer.scala:312) > > at kafka.network.Processor.run(SocketServer.scala:207) > > at java.lang.Thread.run(Thread.java:662) > > > > > > Also please keep informed that this error did happen several times in my > > test (version 0.6 and 0.7). Could you please help me fix this? > > > > Thanks, > > ~Thai > > >