ektravel commented on code in PR #14269:
URL: https://github.com/apache/druid/pull/14269#discussion_r1197924642


##########
docs/configuration/index.md:
##########
@@ -951,16 +951,16 @@ Issuing a GET request at the same URL will return the 
spec that is currently in
 |`millisToWaitBeforeDeleting`|How long does the Coordinator need to be a 
leader before it can start marking overshadowed segments as unused in metadata 
storage.|900000 (15 mins)|
 |`mergeBytesLimit`|The maximum total uncompressed size in bytes of segments to 
merge.|524288000L|
 |`mergeSegmentsLimit`|The maximum number of segments that can be in a single 
[append task](../ingestion/tasks.md).|100|
-|`maxSegmentsToMove`|The maximum number of segments that can be moved at any 
given time.|5|
+|`maxSegmentsToMove`|The maximum number of segments that can be moved at any 
given time.|100|
 |`useBatchedSegmentSampler`|Deprecated. Boolean flag for whether or not we 
should use the Reservoir Sampling with a reservoir of size k instead of fixed 
size 1 to pick segments to move. This option can be enabled to speed up the 
sampling of segments to be balanced, especially if there is a large number of 
segments in the cluster or if there are too many segments to move.|true|
 |`percentOfSegmentsToConsiderPerMove`|Deprecated. This will eventually be 
phased out by the batched segment sampler. You can enable the batched segment 
sampler now by setting the dynamic Coordinator config, 
`useBatchedSegmentSampler`, to `true`. Note that if you choose to enable the 
batched segment sampler, `percentOfSegmentsToConsiderPerMove` will no longer 
have any effect on balancing. If `useBatchedSegmentSampler == false`, this 
config defines the percentage of the total number of segments in the cluster 
that are considered every time a segment needs to be selected for a move. Druid 
orders servers by available capacity ascending (the least available capacity 
first) and then iterates over the servers. For each server, Druid iterates over 
the segments on the server, considering them for moving. The default config of 
100% means that every segment on every server is a candidate to be moved. This 
should make sense for most small to medium-sized clusters. However, an admin 
may find it 
 preferable to drop this value lower if they don't think that it is worthwhile 
to consider every single segment in the cluster each time it is looking for a 
segment to move.|100|
 |`replicantLifetime`|The maximum number of Coordinator runs for a segment to 
be replicated before we start alerting.|15|
-|`replicationThrottleLimit`|The maximum number of segments that can be 
replicated at one time.|10|
+|`replicationThrottleLimit`|The maximum number of segments that can be in the 
replication queue of a historical tier at any given time.|500|
 |`balancerComputeThreads`|Thread pool size for computing moving cost of 
segments in segment balancing. Consider increasing this if you have a lot of 
segments and moving segments starts to get stuck.|1|
 |`emitBalancingStats`|Boolean flag for whether or not we should emit balancing 
stats. This is an expensive operation.|false|
 |`killDataSourceWhitelist`|List of specific data sources for which kill tasks 
are sent if property `druid.coordinator.kill.on` is true. This can be a list of 
comma-separated data source names or a JSON array.|none|
 |`killPendingSegmentsSkipList`|List of data sources for which pendingSegments 
are _NOT_ cleaned up if property `druid.coordinator.kill.pendingSegments.on` is 
true. This can be a list of comma-separated data sources or a JSON array.|none|
-|`maxSegmentsInNodeLoadingQueue`|The maximum number of segments that could be 
queued for loading to any given server. This parameter could be used to speed 
up segments loading process, especially if there are "slow" nodes in the 
cluster (with low loading speed) or if too much segments scheduled to be 
replicated to some particular node (faster loading could be preferred to better 
segments distribution). Desired value depends on segments loading speed, 
acceptable replication time and number of nodes. Value 1000 could be a start 
point for a rather big cluster. Default value is 100. |100|
+|`maxSegmentsInNodeLoadingQueue`|The maximum number of segments that could be 
queued for loading to any given server. This parameter could be used to speed 
up segments loading process, especially if there are "slow" nodes in the 
cluster (with low loading speed) or if too much segments scheduled to be 
replicated to some particular node (faster loading could be preferred to better 
segments distribution). Desired value depends on segments loading speed, 
acceptable replication time and number of nodes. Value 1000 could be a start 
point for a rather big cluster. |500|

Review Comment:
   ```suggestion
   |`maxSegmentsInNodeLoadingQueue`|The maximum number of segments allowed in a 
loading queue for any given server. Use this parameter to load the segments 
faster—for example, if the cluster contains slow-loading nodes, or if 
there are too many segments to be replicated to a particular node (when faster 
loading is preferred to better segments distribution). Desired value depends on 
the loading speed of segments, acceptable replication time, and number of 
nodes. Value 1000 is a good starting point for a big cluster. |500|
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to