This is an automated email from the ASF dual-hosted git repository.

danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 65dd645b487 [HUDI-6959] Bulk insert as row do not rollback failed 
instant on abort (#9887)
65dd645b487 is described below

commit 65dd645b487a61fbca7e4e4b849d1f2f1ec143f9
Author: StreamingFlames <[email protected]>
AuthorDate: Tue Oct 24 20:36:28 2023 -0500

    [HUDI-6959] Bulk insert as row do not rollback failed instant on abort 
(#9887)
---
 .../java/org/apache/hudi/internal/DataSourceInternalWriterHelper.java  | 1 -
 .../src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala     | 3 +--
 2 files changed, 1 insertion(+), 3 deletions(-)

diff --git 
a/hudi-spark-datasource/hudi-spark-common/src/main/java/org/apache/hudi/internal/DataSourceInternalWriterHelper.java
 
b/hudi-spark-datasource/hudi-spark-common/src/main/java/org/apache/hudi/internal/DataSourceInternalWriterHelper.java
index 4ad6c2066a3..58bb3e4d608 100644
--- 
a/hudi-spark-datasource/hudi-spark-common/src/main/java/org/apache/hudi/internal/DataSourceInternalWriterHelper.java
+++ 
b/hudi-spark-datasource/hudi-spark-common/src/main/java/org/apache/hudi/internal/DataSourceInternalWriterHelper.java
@@ -97,7 +97,6 @@ public class DataSourceInternalWriterHelper {
 
   public void abort() {
     LOG.error("Commit " + instantTime + " aborted ");
-    writeClient.rollback(instantTime);
     writeClient.close();
   }
 
diff --git 
a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala
 
b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala
index 14bc84948c1..8cc107a24fb 100644
--- 
a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala
+++ 
b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala
@@ -1714,8 +1714,7 @@ class TestInsertTable extends HoodieSparkSqlTestBase {
     }
   }
 
-  // [HUDI-6900] TestInsertTable "Test Bulk Insert Into Consistent Hashing 
Bucket Index Table" is failing continuously
-  ignore("Test Bulk Insert Into Consistent Hashing Bucket Index Table") {
+  test("Test Bulk Insert Into Consistent Hashing Bucket Index Table") {
     withSQLConf("hoodie.datasource.write.operation" -> "bulk_insert") {
       Seq("false", "true").foreach { bulkInsertAsRow =>
         withTempDir { tmp =>

Reply via email to