Thanks!

On Tue, Apr 18, 2017, 12:26 AM Eno Thereska <eno.there...@gmail.com> wrote:

> The RC candidate build is here:
> http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/ <
> http://home.apache.org/~gwenshap/kafka-0.10.2.1-rc1/>
>
> Eno
> > On 17 Apr 2017, at 17:20, Mahendra Kariya <mahendra.kar...@go-jek.com>
> wrote:
> >
> > Thanks!
> >
> > In the meantime, is the jar published somewhere on github or as a part of
> > build pipeline?
> >
> > On Mon, Apr 17, 2017 at 9:18 PM, Eno Thereska <eno.there...@gmail.com>
> > wrote:
> >
> >> Not yet, but as soon as 0.10.2 is voted it should be. Hopefully this
> week.
> >>
> >> Eno
> >>> On 17 Apr 2017, at 13:25, Mahendra Kariya <mahendra.kar...@go-jek.com>
> >> wrote:
> >>>
> >>> Are the bug fix releases published to Maven central repo?
> >>>
> >>> On Sat, Apr 1, 2017 at 12:26 PM, Eno Thereska <eno.there...@gmail.com>
> >>> wrote:
> >>>
> >>>> Hi Sachin,
> >>>>
> >>>> In the bug fix release for 0.10.2 (and in trunk) we have now set
> >>>> max.poll.interval to infinite since from our experience with streams
> >> this
> >>>> should not be something that users set: https://github.com/apache/
> >>>> kafka/pull/2770/files <
> https://github.com/apache/kafka/pull/2770/files
> >>> .
> >>>>
> >>>> We're in the process of documenting that change. For now you can
> >> increase
> >>>> the request timeout without worrying about max.poll.interval anymore.
> In
> >>>> fact I'd suggest you also increase max.poll.interval as we've done it
> >> above.
> >>>>
> >>>> Thanks
> >>>> Eno
> >>>>
> >>>>> On 1 Apr 2017, at 03:28, Sachin Mittal <sjmit...@gmail.com> wrote:
> >>>>>
> >>>>> Should this timeout be less than max poll interval value? if yes than
> >>>>> generally speaking what should be the ratio between two or range for
> >> this
> >>>>> timeout value .
> >>>>>
> >>>>> Thanks
> >>>>> Sachin
> >>>>>
> >>>>> On 1 Apr 2017 04:57, "Matthias J. Sax" <matth...@confluent.io>
> wrote:
> >>>>>
> >>>>> Yes, you can increase ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >>>>>
> >>>>>
> >>>>> -Matthias
> >>>>>
> >>>>>
> >>>>> On 3/31/17 11:32 AM, Sachin Mittal wrote:
> >>>>>> Hi,
> >>>>>> So I have added the config ProducerConfig.RETRIES_CONFIG,
> >>>>> Integer.MAX_VALUE
> >>>>>> and the NotLeaderForPartitionException is gone.
> >>>>>>
> >>>>>> However we see a new exception especially under heavy load:
> >>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1]
> >> exception
> >>>>>> caught when producing
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.
> >> RecordCollectorImpl.flush(
> >>>>> RecordCollectorImpl.java:127)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>>>> org.apache.kafka.streams.processor.internals.
> >>>> StreamTask$1.run(StreamTask.
> >>>>> java:76)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.StreamsMetricsImpl.
> >>>>> measureLatencyNs(StreamsMetricsImpl.java:188)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.
> >>>> StreamTask.commit(StreamTask.
> >>>>> java:280)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitOne(
> >>>>> StreamThread.java:787)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.commitAll(
> >>>>> StreamThread.java:774)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>>>> org.apache.kafka.streams.processor.internals.
> >> StreamThread.maybeCommit(
> >>>>> StreamThread.java:749)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> at
> >>>>>> org.apache.kafka.streams.processor.internals.StreamThread.runLoop(
> >>>>> StreamThread.java:671)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]        at
> >>>>>> org.apache.kafka.streams.processor.internals.
> >>>>> StreamThread.run(StreamThread.java:378)
> >>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>> org.apache.kafka.common.errors.TimeoutException: Expiring 1
> record(s)
> >>>> for
> >>>>>> new-part-advice-key-table-changelog-1: 30001 ms has passed since
> last
> >>>>> append
> >>>>>>
> >>>>>> So any idea as why TimeoutException is happening.
> >>>>>> Is this controlled by
> >>>>>> ProducerConfig.REQUEST_TIMEOUT_MS_CONFIG
> >>>>>>
> >>>>>> If yes
> >>>>>> What should the value be set in this given that out consumer
> >>>>>> max.poll.interval.ms is defaul 5 minutes.
> >>>>>>
> >>>>>> Is there any other setting that we should try to avoid such errors
> >> which
> >>>>>> causes stream thread to die.
> >>>>>>
> >>>>>> Thanks
> >>>>>> Sachin
> >>>>>>
> >>>>>>
> >>>>>> On Sun, Mar 26, 2017 at 1:39 AM, Eno Thereska <
> eno.there...@gmail.com
> >>>
> >>>>>> wrote:
> >>>>>>
> >>>>>>> Hi Sachin,
> >>>>>>>
> >>>>>>> Not in this case.
> >>>>>>>
> >>>>>>> Thanks
> >>>>>>> Eno
> >>>>>>>
> >>>>>>>> On Mar 25, 2017, at 6:19 PM, Sachin Mittal <sjmit...@gmail.com>
> >>>> wrote:
> >>>>>>>>
> >>>>>>>> OK.
> >>>>>>>> I will try this out.
> >>>>>>>>
> >>>>>>>> Do I need to change anything for
> >>>>>>>> max.in.flight.requests.per.connection
> >>>>>>>>
> >>>>>>>> Thanks
> >>>>>>>> Sachin
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Sat, Mar 25, 2017 at 10:59 PM, Eno Thereska <
> >>>> eno.there...@gmail.com>
> >>>>>>>> wrote:
> >>>>>>>>
> >>>>>>>>> Hi Sachin,
> >>>>>>>>>
> >>>>>>>>> For this particular error, “org.apache.kafka.common.errors.
> >>>>>>>>> NotLeaderForPartitionException: This server is not the leader for
> >>>> that
> >>>>>>>>> topic-partition.”, could you try setting the number of retries to
> >>>>>>> something
> >>>>>>>>> large like this:
> >>>>>>>>>
> >>>>>>>>> Properties props = new Properties();
> >>>>>>>>> props.put(StreamsConfig.APPLICATION_ID_CONFIG, applicationId);
> >>>>>>>>> ...
> >>>>>>>>> props.put(ProducerConfig.RETRIES_CONFIG, Integer.MAX_VALUE);
> >>>>>>>>>
> >>>>>>>>> This will retry the produce requests and should hopefully solve
> >> your
> >>>>>>>>> immediate problem.
> >>>>>>>>>
> >>>>>>>>> Thanks
> >>>>>>>>> Eno
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On 25/03/2017, 08:35, "Sachin Mittal" <sjmit...@gmail.com>
> wrote:
> >>>>>>>>>
> >>>>>>>>> Hi,
> >>>>>>>>> We have encountered another case of series of errors which I
> would
> >>>>>>> need
> >>>>>>>>> more help in understanding.
> >>>>>>>>>
> >>>>>>>>> In logs we see message like this:
> >>>>>>>>> ERROR 2017-03-25 03:41:40,001 [kafka-producer-network-thread |
> >>>>>>>>> 85-StreamThread-3-producer]:
> >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl
> >> -
> >>>>>>>>> task
> >>>>>>>>> [0_1] Error sending record to topic new-part-advice-key-table-
> >>>>>>> changelog.
> >>>>>>>>> No
> >>>>>>>>> more offsets will be recorded for this task and the exception
> will
> >>>>>>>>> eventually be thrown
> >>>>>>>>>
> >>>>>>>>> then some millisecond later
> >>>>>>>>> ERROR 2017-03-25 03:41:40,149 [StreamThread-3]:
> >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread -
> >>>>>>>>> stream-thread
> >>>>>>>>> [StreamThread-3] Failed while executing StreamTask 0_1 due to
> >> flush
> >>>>>>>>> state:
> >>>>>>>>> org.apache.kafka.streams.errors.StreamsException: task [0_1]
> >>>>>>> exception
> >>>>>>>>> caught when producing
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>>>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>>>> RecordCollectorImpl.flush(RecordCollectorImpl.java:127)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>> StreamTask.flushState(
> >>>>>>>>> StreamTask.java:422)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >> StreamThread$4.apply(
> >>>>>>>>> StreamThread.java:555)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>>>>>> performOnAllTasks(StreamThread.java:513)
> >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>>>> StreamThread.flushAllState(StreamThread.java:551)
> >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.StreamThread.
> >>>>>>>>> shutdownTasksAndState(StreamThread.java:463)
> >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>> StreamThread.shutdown(
> >>>>>>>>> StreamThread.java:408)
> >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>>>> StreamThread.run(StreamThread.java:389)
> >>>>>>>>> [kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>> org.apache.kafka.common.errors.NotLeaderForPartitionException:
> >>>> This
> >>>>>>>>> server
> >>>>>>>>> is not the leader for that topic-partition.
> >>>>>>>>>
> >>>>>>>>> finally we get this
> >>>>>>>>> ERROR 2017-03-25 03:41:45,724 [StreamThread-3]:
> >>>>>>>>> com.advice.TestKafkaAdvice
> >>>>>>>>> - Uncaught exception:
> >>>>>>>>> org.apache.kafka.streams.errors.StreamsException: Exception
> >> caught
> >>>>>>> in
> >>>>>>>>> process. taskId=0_1, processor=KSTREAM-SOURCE-0000000000,
> >>>>>>>>> topic=advice-stream, partition=1, offset=48062286
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>>>> StreamTask.process(StreamTask.java:216)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >> StreamThread.runLoop(
> >>>>>>>>> StreamThread.java:651)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>>>> StreamThread.run(StreamThread.java:378)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>> Caused by: org.apache.kafka.streams.errors.StreamsException: task
> >>>>>>>>> [0_1]
> >>>>>>>>> exception caught when producing
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.RecordCollectorImpl.
> >>>>>>>>> checkForException(RecordCollectorImpl.java:119)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>> RecordCollectorImpl.send(
> >>>>>>>>> RecordCollectorImpl.java:76)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>     at
> >>>>>>>>> org.apache.kafka.streams.processor.internals.
> >>>>>>> RecordCollectorImpl.send(
> >>>>>>>>> RecordCollectorImpl.java:64)
> >>>>>>>>> ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na]
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> Again it is not clear why in this case we need to shut down the
> >>>>>>> steams
> >>>>>>>>> thread and eventually the application. Shouldn't we capture this
> >>>>>>> error
> >>>>>>>>> too?
> >>>>>>>>>
> >>>>>>>>> Thanks
> >>>>>>>>> Sachin
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>
> >>>>
> >>>>
> >>
> >>
>
>

Reply via email to