On 23-03-2018 15:06, Thomas SEGISMONT wrote: > Hi Pedro, > > 2018-03-23 13:25 GMT+01:00 Pedro Ruivo <pe...@infinispan.org > <mailto:pe...@infinispan.org>>: > > Hi Thomas, > > Is the test in question using any counter/lock? > > > I have seen the problem on a test for counters, on another one for > locks, as well as well as caches only. > But Vert.x starts the ClusteredLockManager and the CounterManager in all > cases (even if no lock/counter is created/used) > > > I did see similar behavior with the counter's in our server test suite. > The partition handling makes the cache degraded because nodes are > starting and stopping concurrently. > > > As for me I was able to observe the problem even when stopping nodes one > after the other and waiting for cluster to go back to HEALTHY status. > Is it possible that the status of the counter and lock caches are not > taken into account in cluster health?
The counter and lock caches are private. So, they aren't in the cluster health neither their name are returned by getCacheNames() method. > > > I'm not sure if there are any JIRA to tracking. Ryan, Dan do you know? > If there is none, it should be created. > > I improved the counters by making the cache start lazily when you first > get or define a counter [1]. This workaround solved the issue for us. > > As a workaround for your test suite, I suggest to make sure the caches > (___counter_configuration and org.infinispan.LOCK) have finished their > state transfer before stopping the cache managers, by invoking > DefaultCacheManager.getCache(*cache-name*) in all the caches managers. > > Sorry for the inconvenience and the delay in replying. > > > No problem. > > > Cheers, > Pedro > > [1] https://issues.jboss.org/browse/ISPN-8860 > <https://issues.jboss.org/browse/ISPN-8860> > > On 21-03-2018 16:16, Thomas SEGISMONT wrote: > > Hi everyone, > > > > I am working on integrating Infinispan 9.2.Final in vertx-infinispan. > > Before merging I wanted to make sure the test suite passed but it > > doesn't. It's not the always the same test involved. > > > > In the logs, I see a lot of messages like "After merge (or > coordinator > > change), cache still hasn't recovered a majority of members and must > > stay in degraded mode. > > The context involved are "___counter_configuration" and > > "org.infinispan.LOCKS" > > > > Most often it's harmless but, sometimes, I also see this exception > > "ISPN000210: Failed to request state of cache" > > Again the cache involved is either "___counter_configuration" or > > "org.infinispan.LOCKS" > > After this exception, the cache manager is unable to stop. It > blocks in > > method "terminate" (join on cache future). > > > > I thought the test suite was too rough (we stop all nodes at the same > > time). So I changed it to make sure that: > > - nodes start one after the other > > - a new node is started only when the previous one indicates > HEALTHY status > > - nodes stop one after the other > > - a node is stopped only when it indicates HEALTHY status > > Pretty much what we do on Kubernetes for the readiness check > actually. > > But it didn't get any better. > > > > Attached are the logs of such a failing test. > > > > Note that the Vert.x test itself does not fail, it's only when > closing > > nodes that we have issues. > > > > Here's our XML config: > > > > https://github.com/vert-x3/vertx-infinispan/blob/ispn92/src/main/resources/default-infinispan.xml > > <https://github.com/vert-x3/vertx-infinispan/blob/ispn92/src/main/resources/default-infinispan.xml> > > > > Does that ring a bell? Do you need more info? > > > > Regards, > > Thomas > > > > > > > > _______________________________________________ > > infinispan-dev mailing list > > infinispan-dev@lists.jboss.org > <mailto:infinispan-dev@lists.jboss.org> > > https://lists.jboss.org/mailman/listinfo/infinispan-dev > <https://lists.jboss.org/mailman/listinfo/infinispan-dev> > > > _______________________________________________ > infinispan-dev mailing list > infinispan-dev@lists.jboss.org <mailto:infinispan-dev@lists.jboss.org> > https://lists.jboss.org/mailman/listinfo/infinispan-dev > <https://lists.jboss.org/mailman/listinfo/infinispan-dev> > > > > > _______________________________________________ > infinispan-dev mailing list > infinispan-dev@lists.jboss.org > https://lists.jboss.org/mailman/listinfo/infinispan-dev > _______________________________________________ infinispan-dev mailing list infinispan-dev@lists.jboss.org https://lists.jboss.org/mailman/listinfo/infinispan-dev