maytasm commented on PR #14642: URL: https://github.com/apache/druid/pull/14642#issuecomment-1647224419
Druid has a feature where the Coordinator will submit tasks periodically to kill unused segments (this is enable by `druid.coordinator.kill.on`). When this feature is enable, it also has a config `druid.coordinator.kill.maxSegments` which control maximum number of segment to kill per kill task submission. The default value is 100. Just some thoughts I have: - If we already speed up segment nuke by batching the SQL Metadata deleteSegments (https://github.com/apache/druid/pull/14639), do we still need this? - If automatic kill task (`druid.coordinator.kill.on`) is our paved path, should we encourage users to use that (and set the `druid.coordinator.kill.maxSegments`, etc) over manually submitting kill task? If user manually submit kill task, they would not have any guardrail that automatic kill task provide (such as too many segments in one kill task, killing interval that they did not intended, datasource whitelist/blacklist, etc). - Do we need the segment nuke action to return the segment - What happen to the caller of the segment nuke action (i.e. KillTask) if some batches succeeded and some batches failed? - If we do go ahead with this change, I think **not** having a new config is better. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
