How many replicas ,you have for this topic . On Thu, Jun 22, 2017 at 9:19 AM, lk_spark <lk_sp...@163.com> wrote:
> java.lang.IllegalStateException: No current assignment for partition > pages-2 > at org.apache.kafka.clients.consumer.internals.SubscriptionState. > assignedState(SubscriptionState.java:264) > at org.apache.kafka.clients.consumer.internals.SubscriptionState. > needOffsetReset(SubscriptionState.java:336) > at org.apache.kafka.clients.consumer.KafkaConsumer. > seekToEnd(KafkaConsumer.java:1236) > at org.apache.spark.streaming.kafka010.DirectKafkaInputDStream. > latestOffsets(DirectKafkaInputDStream.scala:197) > at org.apache.spark.streaming.kafka010.DirectKafkaInputDStream.compute( > DirectKafkaInputDStream.scala:214) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1$$anonfun$1$$anonfun$apply$7.apply(DStream.scala:341) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1$$anonfun$1$$anonfun$apply$7.apply(DStream.scala:341) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1$$anonfun$1.apply(DStream.scala:340) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1$$anonfun$1.apply(DStream.scala:340) > at org.apache.spark.streaming.dstream.DStream. > createRDDWithLocalProperties(DStream.scala:415) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1.apply(DStream.scala:335) > at org.apache.spark.streaming.dstream.DStream$$anonfun$ > getOrCompute$1.apply(DStream.scala:333) > at scala.Option.orElse(Option.scala:289) > at org.apache.spark.streaming.dstream.DStream.getOrCompute( > DStream.scala:330) > at org.apache.spark.streaming.dstream.ForEachDStream. > generateJob(ForEachDStream.scala:48) > at org.apache.spark.streaming.DStreamGraph$$anonfun$1.apply( > DStreamGraph.scala:117) > at org.apache.spark.streaming.DStreamGraph$$anonfun$1.apply( > DStreamGraph.scala:116) > at scala.collection.TraversableLike$$anonfun$flatMap$1.apply( > TraversableLike.scala:241) > at scala.collection.TraversableLike$$anonfun$flatMap$1.apply( > TraversableLike.scala:241) > at scala.collection.mutable.ResizableArray$class.foreach( > ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at scala.collection.TraversableLike$class.flatMap( > TraversableLike.scala:241) > at scala.collection.AbstractTraversable.flatMap(Traversable.scala:104) > at org.apache.spark.streaming.DStreamGraph.generateJobs( > DStreamGraph.scala:116) > at org.apache.spark.streaming.scheduler.JobGenerator$$ > anonfun$3.apply(JobGenerator.scala:249) > at org.apache.spark.streaming.scheduler.JobGenerator$$ > anonfun$3.apply(JobGenerator.scala:247) > at scala.util.Try$.apply(Try.scala:192) > at org.apache.spark.streaming.scheduler.JobGenerator. > generateJobs(JobGenerator.scala:247) > at org.apache.spark.streaming.scheduler.JobGenerator.org$ > apache$spark$streaming$scheduler$JobGenerator$$processEvent(JobGenerator. > scala:183) > at org.apache.spark.streaming.scheduler.JobGenerator$$anon$ > 1.onReceive(JobGenerator.scala:89) > at org.apache.spark.streaming.scheduler.JobGenerator$$anon$ > 1.onReceive(JobGenerator.scala:88) > at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) > > 2017-06-22 > ------------------------------ > lk_spark > ------------------------------ > > *发件人:*"lk_spark"<lk_sp...@163.com> > *发送时间:*2017-06-22 11:13 > *主题:*spark2.1 kafka0.10 > *收件人:*"user.spark"<user@spark.apache.org> > *抄送:* > > hi,all: > when I run stream application for a few minutes ,I got this error : > > 17/06/22 10:34:56 INFO ConsumerCoordinator: Revoking previously assigned > partitions [comment-0, profile-1, profile-3, cwb-3, bizs-1, cwb-1, > weibocomment-0, bizs-2, pages-0, bizs-4, pages-2, weibo-0, pages-4, > weibo-4, clicks-1, comment-1, weibo-2, clicks-3, weibocomment-4, > weibocomment-2, profile-0, profile-2, profile-4, cwb-4, cwb-2, cwb-0, > bizs-0, bizs-3, pages-1, weibo-1, pages-3, clicks-2, weibo-3, clicks-4, > comment-2, weibocomment-3, clicks-0, weibocomment-1] for group youedata1 > 17/06/22 10:34:56 INFO AbstractCoordinator: (Re-)joining group youedata1 > 17/06/22 10:34:56 INFO AbstractCoordinator: Successfully joined group > youedata1 with generation 3 > 17/06/22 10:34:56 INFO ConsumerCoordinator: Setting newly assigned > partitions [comment-0, profile-1, profile-3, cwb-3, bizs-1, cwb-1, > weibocomment-0, bizs-2, bizs-4, pages-4, weibo-4, clicks-1, comment-1, > clicks-3, weibocomment-4, weibocomment-2, profile-0, profile-2, profile-4, > cwb-4, cwb-2, cwb-0, bizs-0, bizs-3, pages-3, clicks-2, weibo-3, clicks-4, > comment-2, weibocomment-3, clicks-0, weibocomment-1] for group youedata1 > 17/06/22 10:34:56 ERROR JobScheduler: Error generating jobs for time > 1498098896000 ms > java.lang.IllegalStateException: No current assignment for partition > pages-2 > > I don't know why ? > > 2017-06-22 > ------------------------------ > lk_spark > >