[
https://issues.apache.org/jira/browse/HUDI-7611?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Y Ethan Guo updated HUDI-7611:
------------------------------
Status: In Progress (was: Open)
> DELETE operation does not route preCombine/ordering field values to the
> delete records
> --------------------------------------------------------------------------------------
>
> Key: HUDI-7611
> URL: https://issues.apache.org/jira/browse/HUDI-7611
> Project: Apache Hudi
> Issue Type: Bug
> Components: spark
> Reporter: Jonathan Vexler
> Assignee: Y Ethan Guo
> Priority: Critical
> Labels: starter
> Fix For: 0.16.0, 1.0.0
>
>
> org.apache.hudi.client.BaseHoodieWriteClient#delete is defined as:
> {code:java}
> /**
> * Deletes a list of {@link HoodieKey}s from the Hoodie table, at the
> supplied instantTime {@link HoodieKey}s will be
> * de-duped and non-existent keys will be removed before deleting.
> *
> * @param keys {@link List} of {@link HoodieKey}s to be deleted
> * @param instantTime Commit time handle
> * @return Collection of WriteStatus to inspect errors and counts
> */
> public abstract O delete(K keys, final String instantTime); {code}
> It does not take in precombine, only the keys. Therefore, when we write to
> the log block, the precombine is not saved. This api will need to be
> modified, as well as org.apache.hudi.table.HoodieTable#delete,
> org.apache.hudi.table.action.deltacommit.SparkDeleteDeltaCommitActionExecutor#SparkDeleteDeltaCommitActionExecutor
> Also take a look at delete prepped, and see if that needs to be modified as
> well
--
This message was sent by Atlassian Jira
(v8.20.10#820010)