Yeah, that's what I thought. We need to redefine not just restart. Thanks for the info!
I do see the usage of subscribe[K,V] in your DStreams example. Looks simple but its not very obvious how it works :-) I'll watch out for the docs and ScalaDoc. Srikanth On Fri, Jul 22, 2016 at 2:15 PM, Cody Koeninger <c...@koeninger.org> wrote: > No, restarting from a checkpoint won't do it, you need to re-define the > stream. > > Here's the jira for the 0.10 integration > > https://issues.apache.org/jira/browse/SPARK-12177 > > I haven't gotten docs completed yet, but there are examples at > > https://github.com/koeninger/kafka-exactly-once/tree/kafka-0.10 > > On Fri, Jul 22, 2016 at 1:05 PM, Srikanth <srikanth...@gmail.com> wrote: > > In Spark 1.x, if we restart from a checkpoint, will it read from new > > partitions? > > > > If you can, pls point us to some doc/link that talks about Kafka 0.10 > integ > > in Spark 2.0. > > > > On Fri, Jul 22, 2016 at 1:33 PM, Cody Koeninger <c...@koeninger.org> > wrote: > >> > >> For the integration for kafka 0.8, you are literally starting a > >> streaming job against a fixed set of topicapartitions, It will not > >> change throughout the job, so you'll need to restart the spark job if > >> you change kafka partitions. > >> > >> For the integration for kafka 0.10 / spark 2.0, if you use subscribe > >> or subscribepattern, it should pick up new partitions as they are > >> added. > >> > >> On Fri, Jul 22, 2016 at 11:29 AM, Srikanth <srikanth...@gmail.com> > wrote: > >> > Hello, > >> > > >> > I'd like to understand how Spark Streaming(direct) would handle Kafka > >> > partition addition? > >> > Will a running job be aware of new partitions and read from it? > >> > Since it uses Kafka APIs to query offsets and offsets are handled > >> > internally. > >> > > >> > Srikanth > > > > >