[ https://issues.apache.org/jira/browse/IGNITE-15068?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Tanmay Ambre updated IGNITE-15068: ---------------------------------- Description: Hi, We have a 18 node (server) cluster for Ignite. We have multiple client nodes that connect to ignite and use the cache api to get and put data. One of the client node is a Kafka streams application. This application receives an event and for each event: # we lookup the data using the key in ignite. i.e. cache.get(key) # we update the data if we find an entry in ignite i.e. cache.put(key, value) # we insert the data if we don't find an entry in ignite i.e. cache.put(key, value) This application processes more than 30 million events per day. In some scenarios we get multiple events for the same key "consecutively" i.e. the time difference between consecutive events is not more than 10 to 20 milliseconds. We are sure they are processed sequentially as they go to the same Kafka partition. What we have observed is sometimes, the step #1 gives us an old copy of the data (not the one which was updated as part of #2). for e.g. when we get the first event we have following value: Key = 1, value = \{version: 1} when we get the second event for the same key, we update ignite as: Key = 1, value = \{version: 2} //increment version by 1 when we get the third event for the same key and when we lookup the data in ignite instead of getting \{version: 2}, we get {color:#ff0000}{version: 1}{color} also sometimes, when we get the second event - we {color:#ff0000}don't even find the entry in ignite{color} (i.e. \{version:1}) Our caches have the following configuration backups = 1 atomicityMode = ATOMIC writeSynchronizationPolicy = PRIMARY_SYNC readFromBackup = true (default - we are not setting this) Is there something wrong? Should we set "readFromBackup = false"? unfortunately it is very difficult to replicate since it happens just a between 50 to 100 times in a day (i.e. out of 30 million events that we are getting). was: Hi, We have a 18 node (server) cluster for Ignite. We have multiple client nodes that connect to ignite and use the cache api to get and put data. One of the client node is a Kafka streams application. This application receives an event and for each event: # we lookup the data using the key in ignite. i.e. cache.get(key) # we update the data if we find an entry in ignite i.e. cache.put(key, value) # we insert the data if we don't find an entry in ignite i.e. cache.put(key, value) This application processes more than 30 million events per day. In some scenarios we get multiple events for the same key "consecutively" i.e. the time difference between consecutive events is not more than 10 to 20 milliseconds. We are sure they are processed sequentially as they go to the same Kafka partition. What we have observed is sometimes, the step #1 gives us an old copy of the data (not the one which was updated as part of #2). for e.g. when we get the first event for the same key, we have following value: Key = 1, value = \{version: 1} when we get the second even we update ignite as: Key = 1, value = \{version: 2} //increment version by 1 when we get the third event and when we lookup the data in ignite instead of getting \{version: 2}, we get {color:#ff0000}{version: 1}{color} also sometimes, when we get the second event - we {color:#ff0000}don't even find the entry in ignite{color} (i.e. \{version:1}) Our caches have the following configuration backups = 1 atomicityMode = ATOMIC writeSynchronizationPolicy = PRIMARY_SYNC readFromBackup = true (default - we are not setting this) Is there something wrong? Should we set "readFromBackup = false"? unfortunately it is very difficult to replicate since it happens just a between 50 to 100 times in a day (i.e. out of 30 million events that we are getting). > Ignite giving stale data - when data is inserted/updated and immediately read > (from the same client node) > --------------------------------------------------------------------------------------------------------- > > Key: IGNITE-15068 > URL: https://issues.apache.org/jira/browse/IGNITE-15068 > Project: Ignite > Issue Type: Task > Components: cache > Affects Versions: 2.9.1 > Reporter: Tanmay Ambre > Priority: Major > > Hi, > We have a 18 node (server) cluster for Ignite. We have multiple client nodes > that connect to ignite and use the cache api to get and put data. > One of the client node is a Kafka streams application. This application > receives an event and for each event: > # we lookup the data using the key in ignite. i.e. cache.get(key) > # we update the data if we find an entry in ignite i.e. cache.put(key, value) > # we insert the data if we don't find an entry in ignite i.e. cache.put(key, > value) > This application processes more than 30 million events per day. In some > scenarios we get multiple events for the same key "consecutively" i.e. the > time difference between consecutive events is not more than 10 to 20 > milliseconds. We are sure they are processed sequentially as they go to the > same Kafka partition. > What we have observed is sometimes, the step #1 gives us an old copy of the > data (not the one which was updated as part of #2). > > for e.g. when we get the first event we have following value: > Key = 1, value = \{version: 1} > when we get the second event for the same key, we update ignite as: > Key = 1, value = \{version: 2} //increment version by 1 > when we get the third event for the same key and when we lookup the data in > ignite instead of getting \{version: 2}, we get {color:#ff0000}{version: > 1}{color} > also sometimes, when we get the second event - we {color:#ff0000}don't even > find the entry in ignite{color} (i.e. \{version:1}) > Our caches have the following configuration > backups = 1 > atomicityMode = ATOMIC > writeSynchronizationPolicy = PRIMARY_SYNC > readFromBackup = true (default - we are not setting this) > > Is there something wrong? Should we set "readFromBackup = false"? > unfortunately it is very difficult to replicate since it happens just a > between 50 to 100 times in a day (i.e. out of 30 million events that we are > getting). > > -- This message was sent by Atlassian Jira (v8.3.4#803005)