rmdmattingly opened a new pull request, #6067: URL: https://github.com/apache/hbase/pull/6067
When successfully taking an incremental backup, one of our final steps is to delete bulk load metadata from the system table for the bulk loads that needed to be captured in the given backup. This means that we will basically truncate the entire bulk loads system table in a single batch of the deletes after successfully taking an incremental backup. Depending on your usage, one may run tons of bulk loads between backups, so this design is needlessly fragile. We should partition these deletes so that we never erroneously fail a backup due to this; there are a few other cases where we generated unbounded multi requests in the BackupSystemTable that this PR addresses too. A few questions here: 1. Do we want to make the batch size configurable? Seems like yet another config in hbase that I'd like to avoid since it will plausibly never be customized 2. It is okay to just use Table#batch like this, correct? 3. I've tested this code in our QA environment, and unit testing feels like we're just testing the Lists class. That said, would people prefer that I add a unit test? @charlesconnell @ndimiduk -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
