Alexey Kudinkin created HUDI-4992:
-------------------------------------
Summary: Spark Row-writing Bulk Insert produces incorrect Bloom
Filter metadata
Key: HUDI-4992
URL: https://issues.apache.org/jira/browse/HUDI-4992
Project: Apache Hudi
Issue Type: Bug
Affects Versions: 0.12.0
Reporter: Alexey Kudinkin
Assignee: Alexey Kudinkin
Fix For: 0.12.1
Troubleshooting duplicates issue w/ Abhishek Modi from Notion, we've found that
the min/max record key stats are being currently persisted incorrectly into
Parquet metadata, leading to duplicate records being produced in their pipeline
after initial bulk-insert.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)