yihua commented on code in PR #11947:
URL: https://github.com/apache/hudi/pull/11947#discussion_r1802563846


##########
hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/spark/sql/hudi/streaming/HoodieStreamSource.scala:
##########
@@ -71,19 +69,6 @@ class HoodieStreamSource(
     
parameters.get(DataSourceReadOptions.QUERY_TYPE.key).contains(DataSourceReadOptions.QUERY_TYPE_INCREMENTAL_OPT_VAL)
 &&
     
parameters.get(DataSourceReadOptions.INCREMENTAL_FORMAT.key).contains(DataSourceReadOptions.INCREMENTAL_FORMAT_CDC_VAL)
 
-  /**
-   * When hollow commits are found while doing streaming read , unlike batch 
incremental query,
-   * we do not use [[HollowCommitHandling.FAIL]] by default, instead we use 
[[HollowCommitHandling.BLOCK]]
-   * to block processing data from going beyond the hollow commits to avoid 
unintentional skip.
-   *
-   * Users can set 
[[DataSourceReadOptions.INCREMENTAL_READ_HANDLE_HOLLOW_COMMIT]] to
-   * [[HollowCommitHandling.USE_TRANSITION_TIME]] to avoid the blocking 
behavior.
-   */
-  private val hollowCommitHandling: HollowCommitHandling =

Review Comment:
   We discussed this issue and agreed that it's simpler to deprecate the hollow 
commit handling in Hudi 1.0 as the completion time-based incremental query can 
properly support serializability; user should upgrade the table to Hudi 1.0 for 
the correct behavior.
   
   Regarding this PR, there are many places that need clean-up on the hollow 
commit handling.  I'll defer this to a separate PR.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to