leventov commented on a change in pull request #7154: rename maintenance mode 
to decommission
URL: https://github.com/apache/incubator-druid/pull/7154#discussion_r263569607
 
 

 ##########
 File path: docs/content/configuration/index.md
 ##########
 @@ -803,9 +803,9 @@ Issuing a GET request at the same URL will return the spec 
that is currently in
 |`killDataSourceWhitelist`|List of dataSources for which kill tasks are sent 
if property `druid.coordinator.kill.on` is true. This can be a list of 
comma-separated dataSources or a JSON array.|none|
 |`killAllDataSources`|Send kill tasks for ALL dataSources if property 
`druid.coordinator.kill.on` is true. If this is set to true then 
`killDataSourceWhitelist` must not be specified or be empty list.|false|
 |`killPendingSegmentsSkipList`|List of dataSources for which pendingSegments 
are _NOT_ cleaned up if property `druid.coordinator.kill.pendingSegments.on` is 
true. This can be a list of comma-separated dataSources or a JSON array.|none|
-|`maxSegmentsInNodeLoadingQueue`|The maximum number of segments that could be 
queued for loading to any given server. This parameter could be used to speed 
up segments loading process, especially if there are "slow" processes in the 
cluster (with low loading speed) or if too much segments scheduled to be 
replicated to some particular node (faster loading could be preferred to better 
segments distribution). Desired value depends on segments loading speed, 
acceptable replication time and number of processes. Value 1000 could be a 
start point for a rather big cluster. Default value is 0 (loading queue is 
unbounded) |0|
-|`historicalNodesInMaintenance`| List of Historical nodes in maintenance mode. 
Coordinator doesn't assign new segments on those nodes and moves segments from 
the nodes according to a specified priority.|none|
-|`nodesInMaintenancePriority`| Priority of segments from servers in 
maintenance. Coordinator takes ceil(maxSegmentsToMove * (priority / 10)) from 
servers in maitenance during balancing phase, i.e.:<br>0 - no segments from 
servers in maintenance will be processed during balancing<br>5 - 50% segments 
from servers in maintenance<br>10 - 100% segments from servers in 
maintenance<br>By leveraging the priority an operator can prevent general nodes 
from overload or decrease maitenance time instead.|7|
+|`maxSegmentsInNodeLoadingQueue`|The maximum number of segments that could be 
queued for loading to any given server. This parameter could be used to speed 
up segments loading process, especially if there are "slow" nodes in the 
cluster (with low loading speed) or if too much segments scheduled to be 
replicated to some particular node (faster loading could be preferred to better 
segments distribution). Desired value depends on segments loading speed, 
acceptable replication time and number of nodes. Value 1000 could be a start 
point for a rather big cluster. Default value is 0 (loading queue is unbounded) 
|0|
+|`decommissioningNodes`| List of historical servers to 'decommission'. 
Coordinator will not assign new segments to 'decomissioning' servers,  and 
segments will be moved away from them to be placed on 'active' servers at the 
maximum rate specified by `decommissioningMaxPercentOfMaxSegmentsToMove`.|none|
 
 Review comment:
   Two nits:
    - "servers,  and" - double space
    - active is no a term, so I would not put it into quotes. Perhaps better to 
say "non-decommissioning" here.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to