Let me loop in the Ignite dev list as long as I've not heard about such an
issue. Personally, don't see any misconfiguration in your Ignite config.

-
Denis


On Thu, Dec 26, 2019 at 10:17 AM Prasad Bhalerao <
prasadbhalerao1...@gmail.com> wrote:

> I used cache.remove(key) method to delete an entry from cache.
>
> Basically I  was not getting the consistent result on subsequent  API
> calls with the same input.
>
> So I used grid gain console to query the cache. I executed the SQL on
> single node at a time.
> While doing this I found data only on node n1. But same entry was not
> present on nodes n2,n3,n4.
>
> Thanks,
> Prasad
>
>
>
>
> On Thu 26 Dec, 2019, 11:09 PM Denis Magda <dma...@apache.org wrote:
>
>> Hello Prasad,
>>
>> What APIs did you use to remove the entry from the cache and what method
>> did you use to confirm that the entry still exists on some of the nodes?
>>
>> -
>> Denis
>>
>>
>> On Thu, Dec 26, 2019 at 8:54 AM Prasad Bhalerao <
>> prasadbhalerao1...@gmail.com> wrote:
>>
>>> Hi,
>>>
>>> I am using ignite 2.6.0 version and the time out settings are as follows.
>>>
>>> IgniteConfiguration cfg = new IgniteConfiguration();
>>> cfg.setFailureDetectionTimeout(120000);
>>> cfg.setNetworkTimeout(10000);
>>> cfg.setClientFailureDetectionTimeout(120000);
>>>
>>> I have 4 server nodes (n1,n2,n3,n4) and 6 client nodes. I am using a
>>> replicated cache and cache configuration is as shown below.
>>> As you can see write-through is false, read through is true and write
>>> synchronization mode is FULL_SYNC.
>>>
>>> I got an issue, a network entry was removed from network cache but some
>>> how it was removed from only 3 server nodes (n2,n3,n4). I was able to see
>>> the network entry on node n1 consistently for a day(when it was removed).
>>> So I checked the logs for any errors/warnings but I could not find any.
>>> I did not see any segmentation issue in logs, looked like cluster was in
>>> healthy state.
>>> When I checked the cache after 2 days, I could not find that entry.
>>> Cache was in a state as it was supposed to be.  Servers were  not stopped
>>> and restarted during this whole time.
>>>
>>> Some how I am not able to reproduce this issue on dev env.
>>>
>>> Is there any way to investigate/debug this issue? Can someone please
>>> advise?
>>>
>>> private CacheConfiguration networkCacheCfg() {
>>>   CacheConfiguration networkCacheCfg = new 
>>> CacheConfiguration<>(CacheName.NETWORK_CACHE.name());
>>>   networkCacheCfg.setAtomicityMode(CacheAtomicityMode.ATOMIC);
>>>   networkCacheCfg.setWriteThrough(false);
>>>   networkCacheCfg.setReadThrough(true);
>>>   networkCacheCfg.setRebalanceMode(CacheRebalanceMode.ASYNC);
>>>   
>>> networkCacheCfg.setWriteSynchronizationMode(CacheWriteSynchronizationMode.FULL_SYNC);
>>>   networkCacheCfg.setBackups(this.backupCount);
>>>   networkCacheCfg.setCacheMode(CacheMode.REPLICATED);
>>>   Factory<NetworkDataCacheLoader> storeFactory = 
>>> FactoryBuilder.factoryOf(NetworkDataCacheLoader.class);
>>>   networkCacheCfg.setCacheStoreFactory(storeFactory);
>>>   networkCacheCfg.setIndexedTypes(DefaultDataAffinityKey.class, 
>>> NetworkData.class);
>>>   networkCacheCfg.setSqlIndexMaxInlineSize(65);
>>>   RendezvousAffinityFunction affinityFunction = new 
>>> RendezvousAffinityFunction();
>>>   affinityFunction.setExcludeNeighbors(true);
>>>   networkCacheCfg.setAffinity(affinityFunction);
>>>   networkCacheCfg.setStatisticsEnabled(true);
>>>
>>>   return networkCacheCfg;
>>> }
>>>
>>>
>>>
>>> Thanks,
>>> PRasad
>>>
>>>

Reply via email to