I tried our test suite on a slower machine (iMac from 2011). It passes consistently there.
On my laptop, I keep seeing this from time to time (in different tests): 2018-04-19T19:53:09.513 WARN [Context=org.infinispan.LOCKS]ISPN000320: After merge (or coordinator change), cache still hasn't recovered a majority of members and must stay in degraded mode. Current members are [sombrero-19385], lost members are [sombrero-42917], stable members are [sombrero-42917, sombrero-19385] It happens when we shutdown nodes one after the other (even when waiting for cluster status to be "healthy" plus extra 2 seconds). After that the nodes remains blocked in DefaultCacheManager.stop 2018-04-19T19:49:29.242 AVERTISSEMENT Thread Thread[vert.x-worker-thread-5,5,main] has been blocked for 60774 ms, time limit is 60000 io.vertx.core.VertxException: Thread blocked at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1693) at java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323) at java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1729) at java.util.concurrent.CompletableFuture.join(CompletableFuture.java:1934) at org.infinispan.manager.DefaultCacheManager.terminate(DefaultCacheManager.java:688) at org.infinispan.manager.DefaultCacheManager.stopCaches(DefaultCacheManager.java:734) at org.infinispan.manager.DefaultCacheManager.stop(DefaultCacheManager.java:711) at io.vertx.ext.cluster.infinispan.InfinispanClusterManager.lambda$leave$5(InfinispanClusterManager.java:285) at io.vertx.ext.cluster.infinispan.InfinispanClusterManager$$Lambda$421/578931659.handle(Unknown Source) at io.vertx.core.impl.ContextImpl.lambda$executeBlocking$1(ContextImpl.java:265) at io.vertx.core.impl.ContextImpl$$Lambda$27/1330754528.run(Unknown Source) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748) 2018-04-18 17:00 GMT+02:00 Thomas SEGISMONT <tsegism...@gmail.com>: > So here's the Circular Referenced Suppressed Exception > > [stateTransferExecutor-thread--p221-t33] 2018-04-18T16:15:06.662+02:00 > WARN [org.infinispan.statetransfer.InboundTransferTask] ISPN000210: > Failed to request state of cache __vertx.subs from node sombrero-25286, > segments {0} > org.infinispan.remoting.transport.jgroups.SuspectException: ISPN000400: > Node sombrero-25286 was suspected > at org.infinispan.remoting.transport.ResponseCollectors. > remoteNodeSuspected(ResponseCollectors.java:33) > at org.infinispan.remoting.transport.impl.SingleResponseCollector. > targetNotFound(SingleResponseCollector.java:31) > at org.infinispan.remoting.transport.impl.SingleResponseCollector. > targetNotFound(SingleResponseCollector.java:17) > at org.infinispan.remoting.transport.ValidSingleResponseCollector. > addResponse(ValidSingleResponseCollector.java:23) > at org.infinispan.remoting.transport.impl.SingleTargetRequest. > receiveResponse(SingleTargetRequest.java:51) > at org.infinispan.remoting.transport.impl. > SingleTargetRequest.onNewView(SingleTargetRequest.java:42) > at org.infinispan.remoting.transport.jgroups. > JGroupsTransport.addRequest(JGroupsTransport.java:921) > at org.infinispan.remoting.transport.jgroups.JGroupsTransport. > invokeCommand(JGroupsTransport.java:815) > at org.infinispan.remoting.transport.jgroups.JGroupsTransport. > invokeCommand(JGroupsTransport.java:123) > at org.infinispan.remoting.rpc.RpcManagerImpl.invokeCommand( > RpcManagerImpl.java:138) > at org.infinispan.statetransfer.InboundTransferTask.startTransfer( > InboundTransferTask.java:134) > at org.infinispan.statetransfer.InboundTransferTask.requestSegments( > InboundTransferTask.java:113) > at org.infinispan.conflict.impl.StateReceiverImpl$ > SegmentRequest.lambda$requestState$2(StateReceiverImpl.java:164) > at org.infinispan.executors.LimitedExecutor.lambda$executeAsync$1( > LimitedExecutor.java:101) > at org.infinispan.executors.LimitedExecutor.runTasks( > LimitedExecutor.java:144) > at org.infinispan.executors.LimitedExecutor.access$100( > LimitedExecutor.java:33) > at org.infinispan.executors.LimitedExecutor$Runner.run( > LimitedExecutor.java:174) > at java.util.concurrent.ThreadPoolExecutor.runWorker( > ThreadPoolExecutor.java:1149) > at java.util.concurrent.ThreadPoolExecutor$Worker.run( > ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Suppressed: java.util.concurrent.ExecutionException: > org.infinispan.remoting.transport.jgroups.SuspectException: ISPN000400: > Node sombrero-25286 was suspected > at java.util.concurrent.CompletableFuture.reportGet( > CompletableFuture.java:357) > at java.util.concurrent.CompletableFuture.get( > CompletableFuture.java:1915) > at org.infinispan.util.concurrent.CompletableFutures. > await(CompletableFutures.java:82) > at org.infinispan.remoting.rpc.RpcManagerImpl.blocking( > RpcManagerImpl.java:260) > ... 10 more > [CIRCULAR > REFERENCE:org.infinispan.remoting.transport.jgroups.SuspectException: > ISPN000400: Node sombrero-25286 was suspected] > > It does not happen with 9.2.0.Final and prevents from using ISPN embedded > with logback. Do you want me to file an issue ? > > 2018-04-18 11:45 GMT+02:00 Thomas SEGISMONT <tsegism...@gmail.com>: > >> Hi folks, >> >> Sorry I've been busy on other things and couldn't get back to you earlier. >> >> I tried running vertx-infinispan test suite with 9.2.1.Final today. There >> are some problems still but I can't say which ones yet because I hit: >> https://jira.qos.ch/browse/LOGBACK-1027 >> >> We use logback for test logs and all I get is: >> >> 2018-04-18 11:37:46,678 [stateTransferExecutor-thread--p4453-t24] ERROR >> o.i.executors.LimitedExecutor - Exception in task >> java.lang.StackOverflowError: null >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:54) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:60) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:72) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:60) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:72) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:60) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:72) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:60) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:72) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:60) >> at ch.qos.logback.classic.spi.ThrowableProxy.<init>(ThrowablePr >> oxy.java:72) >> ... so on so forth >> >> I will run the suite again without logback and tell you what the actual >> problem is. >> >> Regards, >> Thomas >> >> 2018-03-27 11:15 GMT+02:00 Pedro Ruivo <pe...@infinispan.org>: >> >>> JIRA: https://issues.jboss.org/browse/ISPN-8994 >>> >>> On 27-03-2018 10:08, Pedro Ruivo wrote: >>> > >>> > >>> > On 27-03-2018 09:03, Sebastian Laskawiec wrote: >>> >> At the moment, the cluster health status checker enumerates all >>> caches >>> >> in the cache manager [1] and checks whether those cashes are running >>> >> and not in degraded more [2]. >>> >> >>> >> I'm not sure how counter caches have been implemented. One thing is >>> >> for sure - they should be taken into account in this loop [3]. >>> > >>> > The private caches aren't listed by CacheManager.getCacheNames(). We >>> > have to check them via InternalCacheRegistry.getInternalCacheNames(). >>> > >>> > I'll open a JIRA if you don't mind :) >>> > >>> >> >>> >> [1] >>> >> https://github.com/infinispan/infinispan/blob/master/core/sr >>> c/main/java/org/infinispan/health/impl/ClusterHealthImpl.java#L22 >>> >> >>> >> [2] >>> >> https://github.com/infinispan/infinispan/blob/master/core/sr >>> c/main/java/org/infinispan/health/impl/CacheHealthImpl.java#L25 >>> >> >>> >> [3] >>> >> https://github.com/infinispan/infinispan/blob/master/core/sr >>> c/main/java/org/infinispan/health/impl/ClusterHealthImpl.java#L23-L24 >>> _______________________________________________ >>> infinispan-dev mailing list >>> infinispan-dev@lists.jboss.org >>> https://lists.jboss.org/mailman/listinfo/infinispan-dev >>> >> >> >
_______________________________________________ infinispan-dev mailing list infinispan-dev@lists.jboss.org https://lists.jboss.org/mailman/listinfo/infinispan-dev