Hello

I experimenting an issue with my beam pipeline

I have a pipeline in which I split the work into different branches, then I
do a join using CoGroupByKey, each message has its own unique Key.

For the Join, I used a Session Window, and discarding the messages after
trigger.

I'm using Flink Runner and deployed a KInesis application. But I'm
experiencing  an unbounded growth of the checkpoint data size. When I see
in Flink console, the  following task has the largest checkpoint

join_results/GBK -> ToGBKResult ->
join_results/ConstructCoGbkResultFn/ParMultiDo(ConstructCoGbkResult) -> V


Any Advice ?

Thank you very much!

Reply via email to