n3nash commented on a change in pull request #2612:
URL: https://github.com/apache/hudi/pull/2612#discussion_r585245488



##########
File path: docs/_posts/2021-03-01-hudi-file-sizing.md
##########
@@ -0,0 +1,96 @@
+---
+title: "Apache Hudi File Sizing"
+excerpt: "How Apache hudi manages to maintain optimum sized files to maintain 
read SLAs"
+author: shivnarayan
+category: blog
+---
+
+Apache Hudi is a data platform technology that provides several 
functionalities needed to build and manage a data lake. 
+One of the key features that hudi provides for you is self managing file 
sizing so that users don’t need to worry about 
+small files in their dataset. Having a lot of small files will make it harder 
to maintain your SLAs for read queries. 
+But for streaming data lake use-cases, inherently ingests are going to end up 
having smaller volume of writes, which 
+might result in lot of small files if no special management is done.
+
+# Apache Hudi file size management
+
+Hudi avoids such small files and always writes properly sized files, taking a 
slight hit on ingestion but guaranteeing 

Review comment:
       Instead of saying a `slight hit on ingestion` may be word it. This block 
can be reworded :
   
   Hudi provides ways to write properly sized files during ingestion to 
guarantee SLAs for your read queries. Common approaches to writing very small 
files and then later stitching them together solve for system scalability 
issues posed by small files but might violate query SLA's by exposing small 
files to them. One can leverage Hudi's clustering feature that’s part of Hudi, 
but if you want to have a) self managed file sizing b) Avoid exposing small 
files to queries, automatic file sizing feature is a savior. 
   
   




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to