[
https://issues.apache.org/jira/browse/IGNITE-17279?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Vyacheslav Koptilin updated IGNITE-17279:
-----------------------------------------
Description:
It seems that a coordinator node does not correctly update node2part mapping
for lost partitions.
{noformat}
[test-runner-#1%distributed.CachePartitionLostAfterSupplierHasLeftTest%][root]
dump partitions state for <default>:
----preload sync futures----
nodeId=b57ca812-416d-40d7-bb4f-271994900000
consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest0 isDone=true
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest1 isDone=true
----rebalance futures----
nodeId=b57ca812-416d-40d7-bb4f-271994900000 isDone=true res=true topVer=null
remaining: {}
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 isDone=true res=false
topVer=AffinityTopologyVersion [topVer=4, minorTopVer=0]
remaining: {}
----partition state----
localNodeId=b57ca812-416d-40d7-bb4f-271994900000
grid=distributed.CachePartitionLostAfterSupplierHasLeftTest0
local part=0 counters=Counter [lwm=200, missed=[], maxApplied=200, hwm=200]
fullSize=200 *state=LOST* reservations=0 isAffNode=true
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 part=0 *state=LOST* isAffNode=true
...
localNodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
grid=distributed.CachePartitionLostAfterSupplierHasLeftTest1
local part=0 counters=Counter [lwm=0, missed=[], maxApplied=0, hwm=0]
fullSize=100 *state=LOST* reservations=0 isAffNode=true
nodeId=b57ca812-416d-40d7-bb4f-271994900000 part=0 *state=OWNING*
isAffNode=true
...
{noformat}
was:
It seems that a coordinator node does not correctly update node2part mapping
for lost partitions.
{noformat}
[test-runner-#1%distributed.CachePartitionLostAfterSupplierHasLeftTest%][root]
dump partitions state for <default>:
----preload sync futures----
nodeId=b57ca812-416d-40d7-bb4f-271994900000
consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest0 isDone=true
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest1 isDone=true
----rebalance futures----
nodeId=b57ca812-416d-40d7-bb4f-271994900000 isDone=true res=true topVer=null
remaining: {}
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 isDone=true res=false
topVer=AffinityTopologyVersion [topVer=4, minorTopVer=0]
remaining: {}
----partition state----
localNodeId=b57ca812-416d-40d7-bb4f-271994900000
grid=distributed.CachePartitionLostAfterSupplierHasLeftTest0
local part=0 counters=Counter [lwm=200, missed=[], maxApplied=200, hwm=200]
fullSize=200 state=LOST reservations=0 isAffNode=true
nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 part=0 state=LOST isAffNode=true
...
localNodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
grid=distributed.CachePartitionLostAfterSupplierHasLeftTest1
local part=0 counters=Counter [lwm=0, missed=[], maxApplied=0, hwm=0]
fullSize=100 state=LOST reservations=0 isAffNode=true
nodeId=b57ca812-416d-40d7-bb4f-271994900000 part=0 state=OWNING isAffNode=true
...
{noformat}
> Mapping of partition states to nodes can erroneously skip lost partitions on
> the coordinator node
> -------------------------------------------------------------------------------------------------
>
> Key: IGNITE-17279
> URL: https://issues.apache.org/jira/browse/IGNITE-17279
> Project: Ignite
> Issue Type: Bug
> Reporter: Vyacheslav Koptilin
> Assignee: Vyacheslav Koptilin
> Priority: Minor
>
> It seems that a coordinator node does not correctly update node2part mapping
> for lost partitions.
> {noformat}
> [test-runner-#1%distributed.CachePartitionLostAfterSupplierHasLeftTest%][root]
> dump partitions state for <default>:
> ----preload sync futures----
> nodeId=b57ca812-416d-40d7-bb4f-271994900000
> consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest0
> isDone=true
> nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
> consistentId=distributed.CachePartitionLostAfterSupplierHasLeftTest1
> isDone=true
> ----rebalance futures----
> nodeId=b57ca812-416d-40d7-bb4f-271994900000 isDone=true res=true topVer=null
> remaining: {}
> nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 isDone=true res=false
> topVer=AffinityTopologyVersion [topVer=4, minorTopVer=0]
> remaining: {}
> ----partition state----
> localNodeId=b57ca812-416d-40d7-bb4f-271994900000
> grid=distributed.CachePartitionLostAfterSupplierHasLeftTest0
> local part=0 counters=Counter [lwm=200, missed=[], maxApplied=200, hwm=200]
> fullSize=200 *state=LOST* reservations=0 isAffNode=true
> nodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001 part=0 *state=LOST*
> isAffNode=true
> ...
> localNodeId=20fdfa4a-ddf6-4229-b25e-38cd8d300001
> grid=distributed.CachePartitionLostAfterSupplierHasLeftTest1
> local part=0 counters=Counter [lwm=0, missed=[], maxApplied=0, hwm=0]
> fullSize=100 *state=LOST* reservations=0 isAffNode=true
> nodeId=b57ca812-416d-40d7-bb4f-271994900000 part=0 *state=OWNING*
> isAffNode=true
> ...
> {noformat}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)