nsivabalan commented on issue #4432:
URL: https://github.com/apache/hudi/issues/4432#issuecomment-1002861656
please re-open if the proposed solution does not work.thanks
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and
nsivabalan commented on issue #4432:
URL: https://github.com/apache/hudi/issues/4432#issuecomment-1002316249
Can you also try setting `hoodie.parquet.block.size` in addition to
parquet.max.file.size.
--
This is an automated message from the Apache Git Service.
To respond to the message,
nsivabalan commented on issue #4432:
URL: https://github.com/apache/hudi/issues/4432#issuecomment-1002165461
with bulk_insert, file sizing is not guaranteed as its closely tied to
flushing of data to disk. So, if you are seeing issues w/ bulk_insert, it is
expected. but if you see it with
nsivabalan commented on issue #4432:
URL: https://github.com/apache/hudi/issues/4432#issuecomment-1001807101
Have filed a tracking jira. Will be looking into it.
https://issues.apache.org/jira/browse/HUDI-3110
--
This is an automated message from the Apache Git Service.
To respond
nsivabalan commented on issue #4432:
URL: https://github.com/apache/hudi/issues/4432#issuecomment-1001762429
yes, hudi by default uses gzip for parquet compression. but you can override
the compression if need be.
--
This is an automated message from the Apache Git Service.
To