Mixed up with Spark version. Seems like the issue is different based on Guillermo last mail.
On Wed, Feb 27, 2019 at 1:16 PM Akshay Bhardwaj < akshay.bhardwaj1...@gmail.com> wrote: > Hi Gabor, > > I guess you are looking at Kafka 2.1 but Guillermo mentioned initially > that they are working with Kafka 1.0 > > Akshay Bhardwaj > +91-97111-33849 > > > On Wed, Feb 27, 2019 at 5:41 PM Gabor Somogyi <gabor.g.somo...@gmail.com> > wrote: > >> Where exactly? In Kafka broker configuration section here it's 10080: >> https://kafka.apache.org/documentation/ >> >> offsets.retention.minutes After a consumer group loses all its consumers >> (i.e. becomes empty) its offsets will be kept for this retention period >> before getting discarded. For standalone consumers (using manual >> assignment), offsets will be expired after the time of last commit plus >> this retention period. int 10080 [1,...] high read-only >> >> On Wed, Feb 27, 2019 at 1:04 PM Guillermo Ortiz <konstt2...@gmail.com> >> wrote: >> >>> I'm going to check the value, but I didn't change it., normally, the >>> process is always running but sometimes I have to restarted to apply some >>> changes. Sometimes it starts from the beginning and others continue for the >>> last offset. >>> >>> El mié., 27 feb. 2019 a las 12:25, Akshay Bhardwaj (< >>> akshay.bhardwaj1...@gmail.com>) escribió: >>> >>>> Hi Gabor, >>>> >>>> I am talking about offset.retention.minutes which is set default as >>>> 1440 (or 24 hours) >>>> >>>> Akshay Bhardwaj >>>> +91-97111-33849 >>>> >>>> >>>> On Wed, Feb 27, 2019 at 4:47 PM Gabor Somogyi < >>>> gabor.g.somo...@gmail.com> wrote: >>>> >>>>> Hi Akshay, >>>>> >>>>> The feature what you've mentioned has a default value of 7 days... >>>>> >>>>> BR, >>>>> G >>>>> >>>>> >>>>> On Wed, Feb 27, 2019 at 7:38 AM Akshay Bhardwaj < >>>>> akshay.bhardwaj1...@gmail.com> wrote: >>>>> >>>>>> Hi Guillermo, >>>>>> >>>>>> What was the interval in between restarting the spark job? As a >>>>>> feature in Kafka, a broker deleted offsets for a consumer group after >>>>>> inactivity of 24 hours. >>>>>> In such a case, the newly started spark streaming job will read >>>>>> offsets from beginning for the same groupId. >>>>>> >>>>>> Akshay Bhardwaj >>>>>> +91-97111-33849 >>>>>> >>>>>> >>>>>> On Thu, Feb 21, 2019 at 9:08 PM Gabor Somogyi < >>>>>> gabor.g.somo...@gmail.com> wrote: >>>>>> >>>>>>> From the info you've provided not much to say. >>>>>>> Maybe you could collect sample app, logs etc, open a jira and we can >>>>>>> take a deeper look at it... >>>>>>> >>>>>>> BR, >>>>>>> G >>>>>>> >>>>>>> >>>>>>> On Thu, Feb 21, 2019 at 4:14 PM Guillermo Ortiz < >>>>>>> konstt2...@gmail.com> wrote: >>>>>>> >>>>>>>> I' working with Spark Streaming 2.0.2 and Kafka 1.0.0 using Direct >>>>>>>> Stream as connector. I consume data from Kafka and autosave the >>>>>>>> offsets. >>>>>>>> I can see Spark doing commits in the logs of the last offsets >>>>>>>> processed, Sometimes I have restarted spark and it starts from the >>>>>>>> beginning, when I'm using the same groupId. >>>>>>>> >>>>>>>> Why could it happen? it only happen rarely. >>>>>>>> >>>>>>>