TheR1sing3un commented on issue #12210:
URL: https://github.com/apache/hudi/issues/12210#issuecomment-2461190696

   > Also the main gains for consistent-hasing is to try to rewrite as less 
data files for re-hashing. Otherwise, you have to rewrie all the existing data 
set(the whole table) which is a cost that unaccepted for many cases.
   
   So how can people who are using simple-bucket today deal with the increasing 
amount of data in their buckets? At present, it seems that only the method of 
deleting the table reconstruction can be solved, but the cost is relatively 
high, if we can support dynamically adjusting number of buckets through 
clustering for each partition, will it be more appropriate?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to