Hi Folks, My Spark application interacts with kafka for getting data through Java Api. I am using Direct Approach (No Receivers) - which use Kafka’s simple consumer API to Read data. So, kafka offsets need to be handles explicitly.
In case of Spark failure i need to save the offset state of kafka for resuming from the failure point. I am saving these points in MongoDB. Please tell he how to initialize Kafka DirectStream with saved offset points. I want to initialize kafka stream in Spark Streaming with required offset points. There is method i gets on web. KafkaUtils.createDirectStream(jssc, String.class, String.class, StringDecoder.class, StringDecoder.class, kafkaParams, topicsSet, fromOffsets, arg8); arg8 - kafka.message.MessageAndMetadata Please tell me how to handle and initialize this. Regards, Deepesh