rmdmattingly opened a new pull request, #6067:
URL: https://github.com/apache/hbase/pull/6067

   When successfully taking an incremental backup, one of our final steps is to 
delete bulk load metadata from the system table for the bulk loads that needed 
to be captured in the given backup. This means that we will basically truncate 
the entire bulk loads system table in a single batch of the deletes after 
successfully taking an incremental backup. Depending on your usage, one may run 
tons of bulk loads between backups, so this design is needlessly fragile. We 
should partition these deletes so that we never erroneously fail a backup due 
to this; there are a few other cases where we generated unbounded multi 
requests in the BackupSystemTable that this PR addresses too.
   
   A few questions here:
   1. Do we want to make the batch size configurable? Seems like yet another 
config in hbase that I'd like to avoid since it will plausibly never be 
customized
   2. It is okay to just use Table#batch like this, correct?
   3. I've tested this code in our QA environment, and unit testing feels like 
we're just testing the Lists class. That said, would people prefer that I add a 
unit test?
   
   @charlesconnell @ndimiduk 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to