Re: Does Flink Kafka connector has max_pending_offsets concept?

2019-06-06 Thread xwang355
Thanks Fabian. This is really helpful. -- Sent from: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/

Re: Does Flink Kafka connector has max_pending_offsets concept?

2019-06-06 Thread Fabian Hueske
Hi Ben, Flink correctly maintains the offsets of all partitions that are read by a Kafka consumer. A checkpoint is only complete when all functions successful checkpoint their state. For a Kafka consumer, this state is the current reading offset. In case of a failure the offsets and the state of

Re: Does Flink Kafka connector has max_pending_offsets concept?

2019-06-05 Thread xwang355
Elias Thanks for your reply. In this case, *When # of Kafka consumers = # of partitions, and I use setParallelism(>1), something like this 'messageSteam.rebalance().map(lamba).setParallelism(3).print()' * If checkpointing is enabled, I assume Flink will commit the offsets in the 'right order'

Re: Does Flink Kafka connector has max_pending_offsets concept?

2019-06-05 Thread Elias Levy
There is no such concept in Flink. Flink tracks offsets in its checkpoints. It can optionally commit offsets to Kafka, but that is only for reporting purposes. If you wish to lower the number of records that get reprocessed in the case of a restart, then you must lower the checkpoint interval.

Does Flink Kafka connector has max_pending_offsets concept?

2019-06-04 Thread wang xuchen
Hi Flink users, When # of Kafka consumers = # of partitions, and I use setParallelism(>1), something like this 'messageSteam.rebalance().map(lamba).setParallelism(3).print()' How do I tune # of outstanding uncommitted offset? Something similar to