What is the exact behaviour you're seeing. It's not 100 % clear to me from the initial message.
Best, Aljoscha > On 4. Jan 2018, at 12:56, Sushil Ks <sushil...@gmail.com> wrote: > > *bump* > > > On Dec 15, 2017 11:22 PM, "Lukasz Cwik" <lc...@google.com > <mailto:lc...@google.com>> wrote: > +dev@beam.apache.org <mailto:dev@beam.apache.org> > > On Thu, Dec 14, 2017 at 11:27 PM, Sushil Ks <sushil...@gmail.com > <mailto:sushil...@gmail.com>> wrote: > Hi Likasz, > I am not sure whether I can reproduce in the DirectRunner, as am > taking retry and checkpoint mechanism of Flink into consideration. In other > words, the issue am facing is, any exception in the operation post GroupByKey > and the pipeline restarts, those particular elements are not being processed > in the next run. > > On Wed, Dec 13, 2017 at 4:01 AM, Lukasz Cwik <lc...@google.com > <mailto:lc...@google.com>> wrote: > That seems incorrect. Please file a JIRA and provide an example + data that > shows the error using the DirectRunner. > > On Tue, Dec 12, 2017 at 2:51 AM, Sushil Ks <sushil...@gmail.com > <mailto:sushil...@gmail.com>> wrote: > Hi, > I am running a fixed window with GroupByKey on FlinkRunner and have > noticed that any exception and restart before the GroupByKey operation the > Kafka consumer is replaying the data from the particular offset, however, > post that any exception occurs and the pipeline restart the Kafka is > consuming from the latest offset. Is this expected? > > Regards, > Sushil Ks > > >