XuQianJin-Stars commented on a change in pull request #4489:
URL: https://github.com/apache/hudi/pull/4489#discussion_r780649867
##########
File path:
hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkDeletePartitionCommitActionExecutor.java
##########
@@ -42,27 +46,49 @@
public class SparkDeletePartitionCommitActionExecutor<T extends
HoodieRecordPayload<T>>
extends SparkInsertOverwriteCommitActionExecutor<T> {
- private List<String> partitions;
+ private final List<String> partitions;
+ private final boolean purge;
public SparkDeletePartitionCommitActionExecutor(HoodieEngineContext context,
HoodieWriteConfig config,
HoodieTable table,
- String instantTime,
List<String> partitions) {
- super(context, config, table, instantTime,null,
WriteOperationType.DELETE_PARTITION);
+ String instantTime,
List<String> partitions,
+ boolean purge) {
+ super(context, config, table, instantTime, null,
WriteOperationType.DELETE_PARTITION);
this.partitions = partitions;
+ this.purge = purge;
}
@Override
public HoodieWriteMetadata<JavaRDD<WriteStatus>> execute() {
- JavaSparkContext jsc = HoodieSparkEngineContext.getSparkContext(context);
- HoodieTimer timer = new HoodieTimer().startTimer();
- Map<String, List<String>> partitionToReplaceFileIds =
jsc.parallelize(partitions, partitions.size()).distinct()
- .mapToPair(partitionPath -> new Tuple2<>(partitionPath,
getAllExistingFileIds(partitionPath))).collectAsMap();
- HoodieWriteMetadata result = new HoodieWriteMetadata();
- result.setPartitionToReplaceFileIds(partitionToReplaceFileIds);
- result.setIndexUpdateDuration(Duration.ofMillis(timer.endTimer()));
+ try {
+ JavaSparkContext jsc = HoodieSparkEngineContext.getSparkContext(context);
+ HoodieTimer timer = new HoodieTimer().startTimer();
+ Map<String, List<String>> partitionToReplaceFileIds =
jsc.parallelize(partitions, partitions.size()).distinct()
+ .mapToPair(partitionPath -> new Tuple2<>(partitionPath,
getAllExistingFileIds(partitionPath))).collectAsMap();
+ HoodieWriteMetadata result = new HoodieWriteMetadata();
+ result.setPartitionToReplaceFileIds(partitionToReplaceFileIds);
+ result.setIndexUpdateDuration(Duration.ofMillis(timer.endTimer()));
- result.setWriteStatuses(jsc.emptyRDD());
- this.saveWorkloadProfileMetadataToInflight(new WorkloadProfile(Pair.of(new
HashMap<>(), new WorkloadStat())), instantTime);
- this.commitOnAutoCommit(result);
- return result;
+ // delete partition's path
+ if (purge) {
+ deletePartitionsPath();
+ }
+
+ result.setWriteStatuses(jsc.emptyRDD());
+ this.saveWorkloadProfileMetadataToInflight(new
WorkloadProfile(Pair.of(new HashMap<>(), new WorkloadStat())), instantTime);
+ this.commitOnAutoCommit(result);
+ return result;
+ } catch (Exception e) {
+ throw new HoodieDeletePartitionException("Failed to execute delete
partitions for commit time " + instantTime, e);
+ }
+ }
+
+ private void deletePartitionsPath() throws IOException {
+ String basePath = table.getMetaClient().getBasePath();
+ for (String partition : partitions) {
+ Path fullPartitionPath = FSUtils.getPartitionPath(basePath, partition);
+ if (table.getMetaClient().getFs().exists(fullPartitionPath)) {
+ table.getMetaClient().getFs().delete(fullPartitionPath, true);
Review comment:
> tricky thing is: can we even roll back purge delete partitions? shall
we make an exception for making delete partition irreversible?
Is this `purge` operation really irreversible, or do we need to put this
soft delete in a specific location for easy rollback?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]