satishkotha commented on a change in pull request #2196:
URL: https://github.com/apache/hudi/pull/2196#discussion_r510350766



##########
File path: hudi-spark/src/main/scala/org/apache/hudi/HoodieSparkSqlWriter.scala
##########
@@ -93,6 +93,11 @@ private[hudi] object HoodieSparkSqlWriter {
       operation = WriteOperationType.INSERT
     }
 
+    // If the mode is Overwrite, should use INSERT_OVERWRITE operation

Review comment:
       I think this won't work. Insert overwrite only updates partitions that 
have records in the dataframe. Other partitions continue to have old data.
   
   But, I like the idea. Maybe we can add additional configuration to 
insert_overwrite to mark old partitions as 'deleted'? This can be done in a way 
to support https://issues.apache.org/jira/browse/HUDI-1350. 




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to