maytasm commented on PR #14642:
URL: https://github.com/apache/druid/pull/14642#issuecomment-1647224419

   Druid has a feature where the Coordinator will submit tasks periodically to 
kill unused segments (this is enable by `druid.coordinator.kill.on`). When this 
feature is enable, it also has a config `druid.coordinator.kill.maxSegments` 
which control maximum number of segment to kill per kill task submission. The 
default value is 100.
   
   Just some thoughts I have:
   - If we already speed up segment nuke by batching the SQL Metadata 
deleteSegments (https://github.com/apache/druid/pull/14639), do we still need 
this?
   - If automatic kill task (`druid.coordinator.kill.on`) is our paved path, 
should we encourage users to use that (and set the 
`druid.coordinator.kill.maxSegments`, etc) over manually submitting kill task? 
If user manually submit kill task, they would not have any guardrail that 
automatic kill task provide (such as too many segments in one kill task, 
killing interval that they did not intended, datasource whitelist/blacklist, 
etc).
   - Do we need the segment nuke action to return the segment
   - What happen to the caller of the segment nuke action (i.e. KillTask) if 
some batches succeeded and some batches failed?
   - If we do go ahead with this change, I think **not** having a new config is 
better.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to