nsivabalan commented on code in PR #7632:
URL: https://github.com/apache/hudi/pull/7632#discussion_r1072765472
##########
hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/DataSourceOptions.scala:
##########
@@ -455,6 +455,15 @@ object DataSourceWriteOptions {
+ "This could introduce the potential issue that the job is
restart(`batch id` is lost) while spark checkpoint write fails, "
+ "causing spark will retry and rewrite the data.")
+ val STREAMING_DISABLE_COMPACTION: ConfigProperty[String] = ConfigProperty
Review Comment:
inline compaction does not makes sense for streaming ingestion.So, only
option users have is to leverage async compaction in a separate thread or
completely disable compaction w/ ingestion process and take up async compaction
by a separate process altogether.
So, given this, not sure how we can deduce this. bcoz, default value for
"hoodie.compact.inline" is false which means is async. Can you help me
understand. def interested to see if we can avoid the new config. we also tried
to follow what deltastreamer does towards this to introduce a top level config.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]