[
https://issues.apache.org/jira/browse/IGNITE-3018?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15261688#comment-15261688
]
Taras Ledkov commented on IGNITE-3018:
--------------------------------------
Dmitriy, because current implementation sorts the list of
hash(partition_number, Node). The list size == nodes_count. Also quantity of
such lists equals partitions count.
So the algorithm sorts partitions_count lists.
> Cache affinity calculation is slow with large nodes number
> ----------------------------------------------------------
>
> Key: IGNITE-3018
> URL: https://issues.apache.org/jira/browse/IGNITE-3018
> Project: Ignite
> Issue Type: Bug
> Components: cache
> Reporter: Semen Boikov
> Assignee: Taras Ledkov
> Priority: Critical
> Fix For: 1.6
>
>
> With large number of cache server nodes (> 200) RendezvousAffinityFunction
> and FairAffinityFunction work pretty slow .
> For RendezvousAffinityFunction.assignPartitions can take hundredes of
> milliseconds, for FairAffinityFunction it can take seconds.
> For RendezvousAffinityFunction most time is spent in MD5 hash calculation and
> nodes list sorting. As optimization we can try to cache {partion, node} MD5
> hash or try another hash function. Also several minor optimizations are
> possible (avoid unncecessary allocations, only one thread local 'get', etc).
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)