Hello! [21:16:59,387][WARNING][jvm-pause-detector-worker][IgniteKernal] Possible too long JVM pause: 13584 milliseconds. [21:16:59,388][INFO][tcp-disco-sock-reader-#6][TcpDiscoverySpi] Finished serving remote node connection [rmtAddr=/172.30.4.64:50487, rmtPort=50487 [21:16:59,398][INFO][tcp-disco-msg-worker-#2][TcpDiscoverySpi] Local node seems to be disconnected from topology (failure detection timeout is reached) [failureDetectionTimeout=10000, connCheckInterval=500]
You have 13,5 seconds of GC but failure detection timeout of 10 seconds. This means nodes will be kicked out of topology. Consider adjusting failureDetectionTimeout so it's considerably longer than your GC pauses. Regards, -- Ilya Kasnacheev пн, 25 мар. 2019 г. в 23:15, newigniter <[email protected]>: > I have a problem where my ignite node goes down often. > I attached the full log I have from last time my node crashed. > What I see from the log is problems with GC(possible long GC pause) and if > I > understand correctly after that some locking happens and after some time > node simple crashed. > > Could someone please take a look and point me in the right direction? > If I have long GC pauses, what is the example of something that could cause > it? > I do have some queries on my ignite node which are "heavy" but I added 10gb > of heap memory to both of my nodes and that is much more than the total > amount of data which is being queried at times when the node goes down. > > I currently use Ignite 2.7. I have 2 nodes replicated cluster with 10 GB of > the heap and 30 GB of non-heap memory per node. > > Tnx. > > ignite-5768a4b3.log > < > http://apache-ignite-users.70518.x6.nabble.com/file/t2228/ignite-5768a4b3.log> > > > > > -- > Sent from: http://apache-ignite-users.70518.x6.nabble.com/ >
