Zhenqiu Huang created HUDI-8881:
-----------------------------------
Summary: Potentail dataloss in Flink hudi sink
Key: HUDI-8881
URL: https://issues.apache.org/jira/browse/HUDI-8881
Project: Apache Hudi
Issue Type: Improvement
Components: flink
Reporter: Zhenqiu Huang
Let's say checkpoint A is completed, then coodinator start to hudi commit. If
there is a write error, then the table will be rolledback and flink job will
fail after the exception is thrown. When the job restart from failure, what
will happen? From my understanding, the time between kafka offset in checkpoint
A -1 and offset in checkpoint A will lose.
we should check whether WriteMetadataEvent contains write failure once the
event is received, if any rollback immediately and throw exception to prevent
checkpoint to complete when there is any write failure.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)