Brute force way to do it might be to just have a separate
streaming-kafka-new-consumer subproject, or something along those lines.

On Fri, Dec 4, 2015 at 3:12 AM, Mario Ds Briggs <mario.bri...@in.ibm.com>
wrote:

> >>
> forcing people on kafka 8.x to upgrade their brokers is questionable.
> <<
>
> I agree and i was more thinking maybe there is a way to support both for a
> period of time (of course means some more code to maintain :-)).
>
>
> thanks
> Mario
>
> [image: Inactive hide details for Cody Koeninger ---04/12/2015 12:15:55
> am---Honestly my feeling on any new API is to wait for a point]Cody
> Koeninger ---04/12/2015 12:15:55 am---Honestly my feeling on any new API is
> to wait for a point release before taking it seriously :)
>
> From: Cody Koeninger <c...@koeninger.org>
> To: Mario Ds Briggs/India/IBM@IBMIN
> Cc: "dev@spark.apache.org" <dev@spark.apache.org>
> Date: 04/12/2015 12:15 am
> Subject: Re: Spark Streaming Kafka - DirectKafkaInputDStream: Using the
> new Kafka Consumer API
> ------------------------------
>
>
>
> Honestly my feeling on any new API is to wait for a point release before
> taking it seriously :)
>
> Auth and encryption seem like the only compelling reason to move, but
> forcing people on kafka 8.x to upgrade their brokers is questionable.
>
> On Thu, Dec 3, 2015 at 11:30 AM, Mario Ds Briggs <
> *mario.bri...@in.ibm.com* <mario.bri...@in.ibm.com>> wrote:
>
>    Hi,
>
>    Wanted to pick Cody's mind on what he thinks about
>    DirectKafkaInputDStream/KafkaRDD internally using the new Kafka consumer
>    API. I know the latter is documented as beta-quality, but yet wanted to
>    know if he sees any blockers as to why shouldn't go there shortly. On my
>    side the consideration is that kafka 0.9.0.0 introduced Authentication and
>    Encryption (beta again) between clients & brokers, but this is available
>    only newer Consumer API's and not in the older Low-level/High-level API's.
>
>    From briefly studying the implementation of
>    DirectKafkaInputDStream/KafkaRDD and new Consumer API, my thinking is that
>    it is possible to support the exact current implementation you have using
>    the new API's.
>    One area that isnt so straightforward was the ctor of KafkaRDD fixes
>    the offsetRange (I did read about the deterministic feature you were after)
>    and i couldnt find a direct method in the new Consumer API to get the
>    current 'latest' offset - however one can do a consumer.seekToEnd() and
>    then call a consumer.position().
>    Of course one other benefit is that the new Consumer API's abstracts
>    away having to deal with finding the leader for a partition, so can get rid
>    of that code
>
>    Would be great to get your thoughts.
>
>    thanks in advance
>    Mario
>
>
>
>

Reply via email to