Re: Leaking nioEventLoopGroup threads
Hi Julian. I would like to submit a fix for this, but the relevant section of the code is tricky for me to understand. Basically, if the PooledPlcDriverManager is used with DefaultNettyPlcConnections, when code calls the "close" method of the PlcConnection, the "close" method of the defaultNettyPlcConnection is intercepted and never called, and the nioEventLoopGroup threads are never closed as your patch corrected. The pooled PlcConnection can go stale or become invalidated by the PLC while it is waiting in the pool, so perhaps the fix is to test if the connection isConnected when it is borrowed from the pool, and then explicitly call the closed method before returning it if the connection is not connected? I would have thought the pooledPlcDriverManager would always test for and return a connected PlcConnection, but I seem to remember another discussion on this topic and the preference was expressed to not always call connect() on the PlcConnection before returning. >From the PooledPlcDriverManager class my potential fix is below (see comment starting "//Call the close method..." @Override public PlcConnection getConnection(String url, PlcAuthentication authentication) throws PlcConnectionException { PoolKey poolKey = poolKeyFactory.getPoolKey(url, authentication); if (LOGGER.isDebugEnabled()) { if (authentication != noPlcAuthentication) { LOGGER.debug("Try to borrow an object for url {} and authentication {}", url, authentication); } else { LOGGER.debug("Try to borrow an object for url {}", url); } } PlcConnection plcConnection; try { plcConnection = keyedObjectPool.borrowObject(poolKey); //Call the close method if the connection has lost connection to close threads that might remain open if (!plcConnection.isConnected()) { plcConnection.close(); } } catch (Exception e) { throw new PlcConnectionException(e); } // Used to invalidate a proxy AtomicBoolean proxyInvalidated = new AtomicBoolean(false); return (PlcConnection) Proxy.newProxyInstance(classLoader, new Class[]{ PlcConnection.class}, (proxy, method, args) -> { if (proxyInvalidated.get()) { throw new IllegalStateException("Proxy not valid anymore"); } if ("close".equals(method.getName())) { LOGGER.debug("close called on {}", plcConnection); proxyInvalidated.set(true); keyedObjectPool.returnObject(poolKey, plcConnection); return null; } else { try { return method.invoke(plcConnection, args); } catch (InvocationTargetException e) { if (e.getCause().getClass() == PlcConnectionException.class) { keyedObjectPool.invalidateObject(poolKey, plcConnection); proxyInvalidated.set(true); } throw e; } } }); } On Tue, Aug 25, 2020 at 10:30 AM Julian Feinauer < j.feina...@pragmaticminds.de> wrote: > Hi Adam, > > > > sorry form y late response and in facgt thanks for your feedback, happy to > hear! > > I will merge my branch and close the issue, no worries. > > > > And you are exactly right, just log another issue fort he Pool and we hope > that someone may find the time to have a look there : ) > > > > Julian > > > > PS.: As you already digged into the code you can try to fix it yourself of > course, I would be more than happy to get a PR from you : ) > > > > *Von: *Adam Rossi > *Datum: *Dienstag, 25. August 2020 um 15:45 > *An: *Julian Feinauer > *Cc: *"dev@plc4x.apache.org" > *Betreff: *Re: Leaking nioEventLoopGroup threads > > > > Juilan, > > > > My apologies - your fix did indeed work. The issue is that > PooledPlcDriverManager does not seem to be calling the close method on the > connection. Switching back to PlcDriverManager from PooledPlcDriverManager > results in your new log comments showing up in the log, and more > importantly no more leaks of the nioEventLoopGroup threads. I have tested > the code in a loop for an hour or so and it is working perfectly. > > > > The PooledPlcDriverManager seems to intercept the close method on the > plcConnection (lines 125 - 130): > > > > if ("close".equals(method.getName())) { > > LOGGER.debug("close called on {}", plcConnection); > > proxyInvalidated.set(true); > > keyedObjectPool.returnObject(poolKey, plcConnection); > > return null; > > } else { > > > > Which makes sense as it is trying to keep active connections pooled. > However, when this connection is again retrieved from the pool it seems the > plcConnection connects again and creates an additional nioEventLoopGroup > thread, which is never closed. > > > > I am new to working with this project and Jira. It seems to me that I > should close the issue I just created as your fix does indeed correct the > original issue, and
Re: Leaking nioEventLoopGroup threads
Also thanks to you, Stefano, indeed for the nice community work you do! Julian Von: Stefano Bossi Datum: Dienstag, 25. August 2020 um 20:20 An: , Adam Rossi , Julian Feinauer Betreff: Re: Leaking nioEventLoopGroup threads Yupiii !!! good news !!! Regards, S. On 25/08/2020 15:45, Adam Rossi wrote: Juilan, My apologies - your fix did indeed work. The issue is that PooledPlcDriverManager does not seem to be calling the close method on the connection. Switching back to PlcDriverManager from PooledPlcDriverManager results in your new log comments showing up in the log, and more importantly no more leaks of the nioEventLoopGroup threads. I have tested the code in a loop for an hour or so and it is working perfectly. The PooledPlcDriverManager seems to intercept the close method on the plcConnection (lines 125 - 130): if ("close".equals(method.getName())) { LOGGER.debug("close called on {}", plcConnection); proxyInvalidated.set(true); keyedObjectPool.returnObject(poolKey, plcConnection); return null; } else { Which makes sense as it is trying to keep active connections pooled. However, when this connection is again retrieved from the pool it seems the plcConnection connects again and creates an additional nioEventLoopGroup thread, which is never closed. I am new to working with this project and Jira. It seems to me that I should close the issue I just created as your fix does indeed correct the original issue, and perhaps open another issue on PooledPlcDriverManager for this thread leak? Regards, Adam On Mon, Aug 24, 2020 at 4:39 AM Julian Feinauer < j.feina...@pragmaticminds.de<mailto:j.feina...@pragmaticminds.de>> wrote: Hi, short feedback. I looked into the code and indeed it seems that we had a bug there which could lead tot he socket leak you described. I pushed a fix in the branch: https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel Would you mind taking a look and testing this with your code @Adam Rossi? Thanks! Juliasn Am 24.08.20, 08:26 schrieb "Julian Feinauer" < j.feina...@pragmaticminds.de<mailto:j.feina...@pragmaticminds.de>>: Perhaps, some related questions: - You are using Linux for your Tests? - Do you close all Connections properly? Normally the `PlcConnection.close()` method should close the EventLoop. Julian Am 24.08.20, 08:23 schrieb "Julian Feinauer" < j.feina...@pragmaticminds.de<mailto:j.feina...@pragmaticminds.de>>: Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" < christofer.d...@c-ware.de<mailto:christofer.d...@c-ware.de>>: Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" <mailto:ac.ro...@gmail.com>: Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait<mailto:java.base@13.0.2/sun.nio.ch.EPoll.wait>(Native Method) java.base@13.0.2<mailto:java.base@13.0.2> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2<mailto:java.base@13.0.2> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch<mailto:java.base@13.0.2/sun.nio.ch> .SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSele
Re: Leaking nioEventLoopGroup threads
Yupiii !!! good news !!! Regards, S. On 25/08/2020 15:45, Adam Rossi wrote: > Juilan, > > My apologies - your fix did indeed work. The issue is that > PooledPlcDriverManager does not seem to be calling the close method on the > connection. Switching back to PlcDriverManager from PooledPlcDriverManager > results in your new log comments showing up in the log, and more > importantly no more leaks of the nioEventLoopGroup threads. I have tested > the code in a loop for an hour or so and it is working perfectly. > > The PooledPlcDriverManager seems to intercept the close method on the > plcConnection (lines 125 - 130): > > if ("close".equals(method.getName())) { > LOGGER.debug("close called on {}", plcConnection); > proxyInvalidated.set(true); > keyedObjectPool.returnObject(poolKey, plcConnection); > return null; > } else { > > Which makes sense as it is trying to keep active connections pooled. > However, when this connection is again retrieved from the pool it seems the > plcConnection connects again and creates an additional nioEventLoopGroup > thread, which is never closed. > > I am new to working with this project and Jira. It seems to me that I > should close the issue I just created as your fix does indeed correct the > original issue, and perhaps open another issue on PooledPlcDriverManager > for this thread leak? > > Regards, Adam > > On Mon, Aug 24, 2020 at 4:39 AM Julian Feinauer < > j.feina...@pragmaticminds.de> wrote: > >> Hi, >> >> short feedback. I looked into the code and indeed it seems that we had a >> bug there which could lead tot he socket leak you described. >> I pushed a fix in the branch: >> >> https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel >> >> Would you mind taking a look and testing this with your code @Adam Rossi? >> >> Thanks! >> Juliasn >> >> Am 24.08.20, 08:26 schrieb "Julian Feinauer" < >> j.feina...@pragmaticminds.de>: >> >> Perhaps, some related questions: >> >> - You are using Linux for your Tests? >> - Do you close all Connections properly? >> Normally the `PlcConnection.close()` method should close the EventLoop. >> >> Julian >> >> Am 24.08.20, 08:23 schrieb "Julian Feinauer" < >> j.feina...@pragmaticminds.de>: >> >> Hi Adam, >> >> I will have a look today! >> >> Do we have a Jira Issue for it already? >> >> Julian >> >> Am 24.08.20, 07:38 schrieb "Christofer Dutz" < >> christofer.d...@c-ware.de>: >> >> Hi Adam, >> >> of course that's unfortunate ... also I will not be able to >> address this issue soon as I have to work on the tasks of my research >> project. >> I have one more month to work on this and I'm months behind >> schedule because I have been doing free support way too much lately. >> >> I really hope Julian will be able to help ... he's way more >> into the details of Netty than I am (cause he's got the book ;-) ) >> >> So Julian? ... it would be super awesome if you could take on >> this issue. >> >> Chris >> >> >> >> Am 24.08.20, 00:17 schrieb "Adam Rossi" : >> >> Thanks, I did test with 0.8.0-SNAPSHOT and see the same >> behavior. In every >> plcConnection a nioEventLoopGroup thread is created and >> does not ever seem >> to be destroyed. >> >> I wish I understood the io.netty.channel.EventLoopGroup >> class better to be >> more helpful here. Would an example program that >> reproduces this thread >> leak be useful? >> >> jconsole shows the thread data as: >> >> Name: nioEventLoopGroup-19-1 >> State: RUNNABLE >> Total blocked: 0 Total waited: 0 >> >> Stack trace: >> java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) >> java.base@13.0.2 >> >> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) >> java.base@13.0.2 >> >> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) >>- locked >> io.netty.channel.nio.SelectedSelectionKeySet@1838f97 >>- locked sun.nio.ch.EPollSelectorImpl@1f49287 >> java.base@13.0.2/sun.nio.ch >> .SelectorImpl.select(SelectorImpl.java:141) >> >> app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) >> >> app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) >> >> app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) >> >> app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) >> >> app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >> >>
Re: Leaking nioEventLoopGroup threads
Hi Adam, sorry form y late response and in facgt thanks for your feedback, happy to hear! I will merge my branch and close the issue, no worries. And you are exactly right, just log another issue fort he Pool and we hope that someone may find the time to have a look there : ) Julian PS.: As you already digged into the code you can try to fix it yourself of course, I would be more than happy to get a PR from you : ) Von: Adam Rossi Datum: Dienstag, 25. August 2020 um 15:45 An: Julian Feinauer Cc: "dev@plc4x.apache.org" Betreff: Re: Leaking nioEventLoopGroup threads Juilan, My apologies - your fix did indeed work. The issue is that PooledPlcDriverManager does not seem to be calling the close method on the connection. Switching back to PlcDriverManager from PooledPlcDriverManager results in your new log comments showing up in the log, and more importantly no more leaks of the nioEventLoopGroup threads. I have tested the code in a loop for an hour or so and it is working perfectly. The PooledPlcDriverManager seems to intercept the close method on the plcConnection (lines 125 - 130): if ("close".equals(method.getName())) { LOGGER.debug("close called on {}", plcConnection); proxyInvalidated.set(true); keyedObjectPool.returnObject(poolKey, plcConnection); return null; } else { Which makes sense as it is trying to keep active connections pooled. However, when this connection is again retrieved from the pool it seems the plcConnection connects again and creates an additional nioEventLoopGroup thread, which is never closed. I am new to working with this project and Jira. It seems to me that I should close the issue I just created as your fix does indeed correct the original issue, and perhaps open another issue on PooledPlcDriverManager for this thread leak? Regards, Adam On Mon, Aug 24, 2020 at 4:39 AM Julian Feinauer mailto:j.feina...@pragmaticminds.de>> wrote: Hi, short feedback. I looked into the code and indeed it seems that we had a bug there which could lead tot he socket leak you described. I pushed a fix in the branch: https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel Would you mind taking a look and testing this with your code @Adam Rossi? Thanks! Juliasn Am 24.08.20, 08:26 schrieb "Julian Feinauer" mailto:j.feina...@pragmaticminds.de>>: Perhaps, some related questions: - You are using Linux for your Tests? - Do you close all Connections properly? Normally the `PlcConnection.close()` method should close the EventLoop. Julian Am 24.08.20, 08:23 schrieb "Julian Feinauer" mailto:j.feina...@pragmaticminds.de>>: Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" mailto:christofer.d...@c-ware.de>>: Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" mailto:ac.ro...@gmail.com>>: Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch<http://sun.nio.ch>.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch<http://sun.nio.ch>.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelecto
Re: Leaking nioEventLoopGroup threads
Juilan, My apologies - your fix did indeed work. The issue is that PooledPlcDriverManager does not seem to be calling the close method on the connection. Switching back to PlcDriverManager from PooledPlcDriverManager results in your new log comments showing up in the log, and more importantly no more leaks of the nioEventLoopGroup threads. I have tested the code in a loop for an hour or so and it is working perfectly. The PooledPlcDriverManager seems to intercept the close method on the plcConnection (lines 125 - 130): if ("close".equals(method.getName())) { LOGGER.debug("close called on {}", plcConnection); proxyInvalidated.set(true); keyedObjectPool.returnObject(poolKey, plcConnection); return null; } else { Which makes sense as it is trying to keep active connections pooled. However, when this connection is again retrieved from the pool it seems the plcConnection connects again and creates an additional nioEventLoopGroup thread, which is never closed. I am new to working with this project and Jira. It seems to me that I should close the issue I just created as your fix does indeed correct the original issue, and perhaps open another issue on PooledPlcDriverManager for this thread leak? Regards, Adam On Mon, Aug 24, 2020 at 4:39 AM Julian Feinauer < j.feina...@pragmaticminds.de> wrote: > Hi, > > short feedback. I looked into the code and indeed it seems that we had a > bug there which could lead tot he socket leak you described. > I pushed a fix in the branch: > > https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel > > Would you mind taking a look and testing this with your code @Adam Rossi? > > Thanks! > Juliasn > > Am 24.08.20, 08:26 schrieb "Julian Feinauer" < > j.feina...@pragmaticminds.de>: > > Perhaps, some related questions: > > - You are using Linux for your Tests? > - Do you close all Connections properly? > Normally the `PlcConnection.close()` method should close the EventLoop. > > Julian > > Am 24.08.20, 08:23 schrieb "Julian Feinauer" < > j.feina...@pragmaticminds.de>: > > Hi Adam, > > I will have a look today! > > Do we have a Jira Issue for it already? > > Julian > > Am 24.08.20, 07:38 schrieb "Christofer Dutz" < > christofer.d...@c-ware.de>: > > Hi Adam, > > of course that's unfortunate ... also I will not be able to > address this issue soon as I have to work on the tasks of my research > project. > I have one more month to work on this and I'm months behind > schedule because I have been doing free support way too much lately. > > I really hope Julian will be able to help ... he's way more > into the details of Netty than I am (cause he's got the book ;-) ) > > So Julian? ... it would be super awesome if you could take on > this issue. > > Chris > > > > Am 24.08.20, 00:17 schrieb "Adam Rossi" : > > Thanks, I did test with 0.8.0-SNAPSHOT and see the same > behavior. In every > plcConnection a nioEventLoopGroup thread is created and > does not ever seem > to be destroyed. > > I wish I understood the io.netty.channel.EventLoopGroup > class better to be > more helpful here. Would an example program that > reproduces this thread > leak be useful? > > jconsole shows the thread data as: > > Name: nioEventLoopGroup-19-1 > State: RUNNABLE > Total blocked: 0 Total waited: 0 > > Stack trace: > java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) > java.base@13.0.2 > > /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) > java.base@13.0.2 > > /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) >- locked > io.netty.channel.nio.SelectedSelectionKeySet@1838f97 >- locked sun.nio.ch.EPollSelectorImpl@1f49287 > java.base@13.0.2/sun.nio.ch > .SelectorImpl.select(SelectorImpl.java:141) > > app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) > > app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) > > app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) > > app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) > > app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > > app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > java.base@13.0.2/java.lang.Thread.run(Thread.java:830) > > > Regards, Adam > > On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz < > christofer.d...@c-ware.de> > wrote: > >
Re: Leaking nioEventLoopGroup threads
Hi Adam, it's much better to not forget your problem if you open a new ticket on Jira. here is the link: https://issues.apache.org/jira/projects/PLC4X/summary You should open an account a copy your mail in there. Just a suggestion to keep the things in order. Thanks, S. On 24/08/2020 17:26, Adam Rossi wrote: > Julian, thank you so much for your attention on this. Your code changes > look good but unfortunately I am still experiencing the problem. > > Here is some log output from my program. You can see that the close is > called on the DefaultNettyPlcConnection, but I do not see the output from > your code modification that I would expect if the closeEventLoopForChannel > method was being called (Either logger.info("Channel is closed, closing > worker Group also") or logger.warn("Trying to remove EventLoop for Channel > {} but have none stored", channel). > > The nioEventLoopGroup threads continue to persist after every > plcConnection. > > I hope I have built everything correctly...I checked out your branch with: > > git clone --single-branch --branch bugfix/close-eventloop-after-channel > https://github.com/apache/plc4x.git > > And wiped out my local m2 maven repository before building the code with: > > ./mvnw install -DskipTests > > I also removed references to the apache snapshot repo from my project pom > and by all appearances I am using the correct 0.8.0-SNAPSHOT jars that are > locally built in my local m2 repo. Here is some log info from my test: > > > > > 2020-08-24_10:15:43.450 DEBUG PooledPlcDriverManager - Try to borrow an > object for url modbus://192.168.0.5:503?unit-identifier=50 > 2020-08-24_10:15:43.452 INFO TcpChannelFactory - Configuring Bootstrap > with Configuration{} > 2020-08-24_10:15:43.458 DEBUG ModbusManager - Connection Metadata: > 2020-08-24_10:15:43.459 DEBUG ModbusManager - > org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection@185c140 > 2020-08-24_10:15:43.462 DEBUG Plc4xNettyWrapper - Forwarding request to plc > ModbusTcpADU[transactionIdentifier=10,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=67,quantity=1]] > 2020-08-24_10:15:43.466 DEBUG GeneratedDriverByteToMessageCodec - Sending > bytes to PLC for message > ModbusTcpADU[transactionIdentifier=10,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=67,quantity=1]] > as data 000a0006320300430001 > 2020-08-24_10:15:43.470 DEBUG Plc4xNettyWrapper - Forwarding request to plc > ModbusTcpADU[transactionIdentifier=11,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=69,quantity=1]] > 2020-08-24_10:15:43.471 DEBUG GeneratedDriverByteToMessageCodec - Sending > bytes to PLC for message > ModbusTcpADU[transactionIdentifier=11,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=69,quantity=1]] > as data 000b0006320300450001 > 2020-08-24_10:15:43.480 DEBUG Plc4xNettyWrapper - Forwarding request to plc > ModbusTcpADU[transactionIdentifier=12,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=68,quantity=1]] > 2020-08-24_10:15:43.481 DEBUG GeneratedDriverByteToMessageCodec - Sending > bytes to PLC for message > ModbusTcpADU[transactionIdentifier=12,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=68,quantity=1]] > as data 000c0006320300440001 > 2020-08-24_10:15:43.484 DEBUG Plc4xNettyWrapper - Forwarding request to plc > ModbusTcpADU[transactionIdentifier=13,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=66,quantity=1]] > 2020-08-24_10:15:43.485 DEBUG GeneratedDriverByteToMessageCodec - Sending > bytes to PLC for message > ModbusTcpADU[transactionIdentifier=13,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=66,quantity=1]] > as data 000d0006320300420001 > 2020-08-24_10:15:43.489 DEBUG PooledPlcDriverManager - close called on > org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection@185c140 > 2020-08-24_10:15:43.490 INFO ReadModbusTask - Read Modbus Task Completed. > > Some details from my code: > > Getting the connection: > > PlcConnection plcConnection = > pooledDriverManager.getConnection(modbusServerURI); > if (plcConnection.isConnected()) { > LOG.trace("The connection is connected"); > } else { > LOG.trace("The connection is not connected. Connecting > now..."); > plcConnection.connect(); > } > > Reading the plc and closing the connection: > > PlcReadRequest.Builder builder = plcConnection.readRequestBuilder(); > builder.addItem("devicename", "holding-register:1[8]"); > builder.addItem("generatorstate", "holding-register:67"); > builder.addItem("batteryvoltage", "holding-register:513[2]"); > builder.addItem("batterycurrent", "holding-register:515[2]"); > builder.addItem("pvpowerwatts", "holding-register:69[2]"); > builder.addItem("pvinputthishourkwh", "holding-register:307[2]"); > PlcReadRequest readRequest = builder.build(); >
Re: Leaking nioEventLoopGroup threads
Julian, thank you so much for your attention on this. Your code changes look good but unfortunately I am still experiencing the problem. Here is some log output from my program. You can see that the close is called on the DefaultNettyPlcConnection, but I do not see the output from your code modification that I would expect if the closeEventLoopForChannel method was being called (Either logger.info("Channel is closed, closing worker Group also") or logger.warn("Trying to remove EventLoop for Channel {} but have none stored", channel). The nioEventLoopGroup threads continue to persist after every plcConnection. I hope I have built everything correctly...I checked out your branch with: git clone --single-branch --branch bugfix/close-eventloop-after-channel https://github.com/apache/plc4x.git And wiped out my local m2 maven repository before building the code with: ./mvnw install -DskipTests I also removed references to the apache snapshot repo from my project pom and by all appearances I am using the correct 0.8.0-SNAPSHOT jars that are locally built in my local m2 repo. Here is some log info from my test: 2020-08-24_10:15:43.450 DEBUG PooledPlcDriverManager - Try to borrow an object for url modbus://192.168.0.5:503?unit-identifier=50 2020-08-24_10:15:43.452 INFO TcpChannelFactory - Configuring Bootstrap with Configuration{} 2020-08-24_10:15:43.458 DEBUG ModbusManager - Connection Metadata: 2020-08-24_10:15:43.459 DEBUG ModbusManager - org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection@185c140 2020-08-24_10:15:43.462 DEBUG Plc4xNettyWrapper - Forwarding request to plc ModbusTcpADU[transactionIdentifier=10,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=67,quantity=1]] 2020-08-24_10:15:43.466 DEBUG GeneratedDriverByteToMessageCodec - Sending bytes to PLC for message ModbusTcpADU[transactionIdentifier=10,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=67,quantity=1]] as data 000a0006320300430001 2020-08-24_10:15:43.470 DEBUG Plc4xNettyWrapper - Forwarding request to plc ModbusTcpADU[transactionIdentifier=11,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=69,quantity=1]] 2020-08-24_10:15:43.471 DEBUG GeneratedDriverByteToMessageCodec - Sending bytes to PLC for message ModbusTcpADU[transactionIdentifier=11,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=69,quantity=1]] as data 000b0006320300450001 2020-08-24_10:15:43.480 DEBUG Plc4xNettyWrapper - Forwarding request to plc ModbusTcpADU[transactionIdentifier=12,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=68,quantity=1]] 2020-08-24_10:15:43.481 DEBUG GeneratedDriverByteToMessageCodec - Sending bytes to PLC for message ModbusTcpADU[transactionIdentifier=12,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=68,quantity=1]] as data 000c0006320300440001 2020-08-24_10:15:43.484 DEBUG Plc4xNettyWrapper - Forwarding request to plc ModbusTcpADU[transactionIdentifier=13,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=66,quantity=1]] 2020-08-24_10:15:43.485 DEBUG GeneratedDriverByteToMessageCodec - Sending bytes to PLC for message ModbusTcpADU[transactionIdentifier=13,unitIdentifier=50,pdu=ModbusPDUReadHoldingRegistersRequest[startingAddress=66,quantity=1]] as data 000d0006320300420001 2020-08-24_10:15:43.489 DEBUG PooledPlcDriverManager - close called on org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection@185c140 2020-08-24_10:15:43.490 INFO ReadModbusTask - Read Modbus Task Completed. Some details from my code: Getting the connection: PlcConnection plcConnection = pooledDriverManager.getConnection(modbusServerURI); if (plcConnection.isConnected()) { LOG.trace("The connection is connected"); } else { LOG.trace("The connection is not connected. Connecting now..."); plcConnection.connect(); } Reading the plc and closing the connection: PlcReadRequest.Builder builder = plcConnection.readRequestBuilder(); builder.addItem("devicename", "holding-register:1[8]"); builder.addItem("generatorstate", "holding-register:67"); builder.addItem("batteryvoltage", "holding-register:513[2]"); builder.addItem("batterycurrent", "holding-register:515[2]"); builder.addItem("pvpowerwatts", "holding-register:69[2]"); builder.addItem("pvinputthishourkwh", "holding-register:307[2]"); PlcReadRequest readRequest = builder.build(); PlcReadResponse response; try { response = readRequest.execute().get(5, TimeUnit.SECONDS); //process response catch (InterruptedException | ExecutionException | TimeoutException ex) { LOG.error("Error in executing gateway read - interrupted or timed out", ex); } } finally { plcConnection.close(); } catch (Exception ex) { LOG.error("Error closing modbus connection", ex);
Re: Leaking nioEventLoopGroup threads
Yes thanks ! This will be very important for my stability test too! Let's see if Adam confirm the patch ! Thanks, Stefano On 24/08/2020 11:35, Christofer Dutz wrote: > Thanks Julian for that quick fix. > > Chris > > > Am 24.08.20, 10:39 schrieb "Julian Feinauer" : > > Hi, > > short feedback. I looked into the code and indeed it seems that we had a > bug there which could lead tot he socket leak you described. > I pushed a fix in the branch: > > https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel > > Would you mind taking a look and testing this with your code @Adam Rossi? > > Thanks! > Juliasn > > Am 24.08.20, 08:26 schrieb "Julian Feinauer" > : > > Perhaps, some related questions: > > - You are using Linux for your Tests? > - Do you close all Connections properly? > Normally the `PlcConnection.close()` method should close the > EventLoop. > > Julian > > Am 24.08.20, 08:23 schrieb "Julian Feinauer" > : > > Hi Adam, > > I will have a look today! > > Do we have a Jira Issue for it already? > > Julian > > Am 24.08.20, 07:38 schrieb "Christofer Dutz" > : > > Hi Adam, > > of course that's unfortunate ... also I will not be able to > address this issue soon as I have to work on the tasks of my research project. > I have one more month to work on this and I'm months behind > schedule because I have been doing free support way too much lately. > > I really hope Julian will be able to help ... he's way more > into the details of Netty than I am (cause he's got the book ;-) ) > > So Julian? ... it would be super awesome if you could take on > this issue. > > Chris > > > > Am 24.08.20, 00:17 schrieb "Adam Rossi" : > > Thanks, I did test with 0.8.0-SNAPSHOT and see the same > behavior. In every > plcConnection a nioEventLoopGroup thread is created and > does not ever seem > to be destroyed. > > I wish I understood the io.netty.channel.EventLoopGroup > class better to be > more helpful here. Would an example program that > reproduces this thread > leak be useful? > > jconsole shows the thread data as: > > Name: nioEventLoopGroup-19-1 > State: RUNNABLE > Total blocked: 0 Total waited: 0 > > Stack trace: > java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) > java.base@13.0.2 > > /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) > java.base@13.0.2 > > /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) >- locked > io.netty.channel.nio.SelectedSelectionKeySet@1838f97 >- locked sun.nio.ch.EPollSelectorImpl@1f49287 > > java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) > > app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) > > app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) > > app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) > > app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) > > app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > > app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > java.base@13.0.2/java.lang.Thread.run(Thread.java:830) > > > Regards, Adam > > On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz > > wrote: > > > Hi Adam, > > > > the Apache SNAPSHOT repo is at: > > > https://repository.apache.org/content/repositories/snapshots > > > > Adding this to your pom should help: > > > > > > > > > > apache-snapshots > > > https://repository.apache.org/content/repositories/snapshots > > > > > > false > > > > > > true > > > > > > > > > > >
Re: Leaking nioEventLoopGroup threads
Thanks Julian for that quick fix. Chris Am 24.08.20, 10:39 schrieb "Julian Feinauer" : Hi, short feedback. I looked into the code and indeed it seems that we had a bug there which could lead tot he socket leak you described. I pushed a fix in the branch: https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel Would you mind taking a look and testing this with your code @Adam Rossi? Thanks! Juliasn Am 24.08.20, 08:26 schrieb "Julian Feinauer" : Perhaps, some related questions: - You are using Linux for your Tests? - Do you close all Connections properly? Normally the `PlcConnection.close()` method should close the EventLoop. Julian Am 24.08.20, 08:23 schrieb "Julian Feinauer" : Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" : Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" : Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > >
Re: Leaking nioEventLoopGroup threads
Hi, short feedback. I looked into the code and indeed it seems that we had a bug there which could lead tot he socket leak you described. I pushed a fix in the branch: https://github.com/apache/plc4x/tree/bugfix/close-eventloop-after-channel Would you mind taking a look and testing this with your code @Adam Rossi? Thanks! Juliasn Am 24.08.20, 08:26 schrieb "Julian Feinauer" : Perhaps, some related questions: - You are using Linux for your Tests? - Do you close all Connections properly? Normally the `PlcConnection.close()` method should close the EventLoop. Julian Am 24.08.20, 08:23 schrieb "Julian Feinauer" : Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" : Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" : Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > Chris > Am 23.08.20, 18:56 schrieb "Adam Rossi" : > > Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something > that needs to be built? Regards Adam > > > On Aug 23, 2020, at 12:31 PM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > >
Re: Leaking nioEventLoopGroup threads
Perhaps, some related questions: - You are using Linux for your Tests? - Do you close all Connections properly? Normally the `PlcConnection.close()` method should close the EventLoop. Julian Am 24.08.20, 08:23 schrieb "Julian Feinauer" : Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" : Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" : Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > Chris > Am 23.08.20, 18:56 schrieb "Adam Rossi" : > > Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something > that needs to be built? Regards Adam > > > On Aug 23, 2020, at 12:31 PM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > > > > Hmm ... > > > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... > 0.7.0 was the first of the new generation drivers. We're maintaining the > 0.6 branch and working hard on making the new generation drivers 100% > production ready. > > > > Chris > > > > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > > > >This is the latest 0.7.0 release from Maven. > > > > > >Regards Adam > > > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz < >
Re: Leaking nioEventLoopGroup threads
Hi Adam, I will have a look today! Do we have a Jira Issue for it already? Julian Am 24.08.20, 07:38 schrieb "Christofer Dutz" : Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" : Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > Chris > Am 23.08.20, 18:56 schrieb "Adam Rossi" : > > Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something > that needs to be built? Regards Adam > > > On Aug 23, 2020, at 12:31 PM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > > > > Hmm ... > > > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... > 0.7.0 was the first of the new generation drivers. We're maintaining the > 0.6 branch and working hard on making the new generation drivers 100% > production ready. > > > > Chris > > > > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > > > >This is the latest 0.7.0 release from Maven. > > > > > >Regards Adam > > > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > >> > >> Hi Adam, > >> > >> which version of PLC4X are you using? I know we had similar reports > some time ago, but had thought we had fixed them > >> > >> Chris > >> > >> > >> > >> Am 23.08.20, 16:40 schrieb "Adam Rossi" : > >> > >> Howdy. I am seeing a persistent thread being created for every > >> plcConnection connect which looks like the following: > >> > >> Name: nioEventLoopGroup-11-1 > >> State: RUNNABLE > >> Total blocked: 0
Re: Leaking nioEventLoopGroup threads
Hi Adam, of course that's unfortunate ... also I will not be able to address this issue soon as I have to work on the tasks of my research project. I have one more month to work on this and I'm months behind schedule because I have been doing free support way too much lately. I really hope Julian will be able to help ... he's way more into the details of Netty than I am (cause he's got the book ;-) ) So Julian? ... it would be super awesome if you could take on this issue. Chris Am 24.08.20, 00:17 schrieb "Adam Rossi" : Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > Chris > Am 23.08.20, 18:56 schrieb "Adam Rossi" : > > Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something > that needs to be built? Regards Adam > > > On Aug 23, 2020, at 12:31 PM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > > > > Hmm ... > > > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... > 0.7.0 was the first of the new generation drivers. We're maintaining the > 0.6 branch and working hard on making the new generation drivers 100% > production ready. > > > > Chris > > > > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > > > >This is the latest 0.7.0 release from Maven. > > > > > >Regards Adam > > > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > >> > >> Hi Adam, > >> > >> which version of PLC4X are you using? I know we had similar reports > some time ago, but had thought we had fixed them > >> > >> Chris > >> > >> > >> > >> Am 23.08.20, 16:40 schrieb "Adam Rossi" : > >> > >> Howdy. I am seeing a persistent thread being created for every > >> plcConnection connect which looks like the following: > >> > >> Name: nioEventLoopGroup-11-1 > >> State: RUNNABLE > >> Total blocked: 0 Total waited: 0 > >> > >> Stack trace: > >> java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) > >> java.base@13.0.2 > >> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) > >> java.base@13.0.2 > >> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) > >> - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b > >> - locked sun.nio.ch.EPollSelectorImpl@7ffc19 > >> java.base@13.0.2/sun.nio.ch > .SelectorImpl.select(SelectorImpl.java:141) > >> >
Re: Leaking nioEventLoopGroup threads
Thanks, I did test with 0.8.0-SNAPSHOT and see the same behavior. In every plcConnection a nioEventLoopGroup thread is created and does not ever seem to be destroyed. I wish I understood the io.netty.channel.EventLoopGroup class better to be more helpful here. Would an example program that reproduces this thread leak be useful? jconsole shows the thread data as: Name: nioEventLoopGroup-19-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@1838f97 - locked sun.nio.ch.EPollSelectorImpl@1f49287 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) Regards, Adam On Sun, Aug 23, 2020 at 1:00 PM Christofer Dutz wrote: > Hi Adam, > > the Apache SNAPSHOT repo is at: > https://repository.apache.org/content/repositories/snapshots > > Adding this to your pom should help: > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > > > > apache-snapshots > https://repository.apache.org/content/repositories/snapshots > > > false > > > true > > > > > Chris > Am 23.08.20, 18:56 schrieb "Adam Rossi" : > > Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something > that needs to be built? Regards Adam > > > On Aug 23, 2020, at 12:31 PM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > > > > Hmm ... > > > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... > 0.7.0 was the first of the new generation drivers. We're maintaining the > 0.6 branch and working hard on making the new generation drivers 100% > production ready. > > > > Chris > > > > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > > > >This is the latest 0.7.0 release from Maven. > > > > > >Regards Adam > > > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz < > christofer.d...@c-ware.de> wrote: > >> > >> Hi Adam, > >> > >> which version of PLC4X are you using? I know we had similar reports > some time ago, but had thought we had fixed them > >> > >> Chris > >> > >> > >> > >> Am 23.08.20, 16:40 schrieb "Adam Rossi" : > >> > >> Howdy. I am seeing a persistent thread being created for every > >> plcConnection connect which looks like the following: > >> > >> Name: nioEventLoopGroup-11-1 > >> State: RUNNABLE > >> Total blocked: 0 Total waited: 0 > >> > >> Stack trace: > >> java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) > >> java.base@13.0.2 > >> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) > >> java.base@13.0.2 > >> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) > >> - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b > >> - locked sun.nio.ch.EPollSelectorImpl@7ffc19 > >> java.base@13.0.2/sun.nio.ch > .SelectorImpl.select(SelectorImpl.java:141) > >> > > app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) > >> > app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) > >> app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) > >> > > app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) > >> > > app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > >> > > app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) > >> java.base@13.0.2/java.lang.Thread.run(Thread.java:830) > >> > >> > >> This happens both when creating a connection directly or when > using the > >> plc4x connection pool. > >> > >> In both cases I am obtaining the connection, reading some modbus > values, > >> and then closing the connection. From the plcConnection metadata > the > >> connection type I am using is: > >> > >>
Re: Leaking nioEventLoopGroup threads
Hi Adam, the Apache SNAPSHOT repo is at: https://repository.apache.org/content/repositories/snapshots Adding this to your pom should help: apache-snapshots https://repository.apache.org/content/repositories/snapshots false true apache-snapshots https://repository.apache.org/content/repositories/snapshots false true Chris Am 23.08.20, 18:56 schrieb "Adam Rossi" : Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something that needs to be built? Regards Adam > On Aug 23, 2020, at 12:31 PM, Christofer Dutz wrote: > > Hmm ... > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... 0.7.0 was the first of the new generation drivers. We're maintaining the 0.6 branch and working hard on making the new generation drivers 100% production ready. > > Chris > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > >This is the latest 0.7.0 release from Maven. > > >Regards Adam > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz wrote: >> >> Hi Adam, >> >> which version of PLC4X are you using? I know we had similar reports some time ago, but had thought we had fixed them >> >> Chris >> >> >> >> Am 23.08.20, 16:40 schrieb "Adam Rossi" : >> >> Howdy. I am seeing a persistent thread being created for every >> plcConnection connect which looks like the following: >> >> Name: nioEventLoopGroup-11-1 >> State: RUNNABLE >> Total blocked: 0 Total waited: 0 >> >> Stack trace: >> java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) >> java.base@13.0.2 >> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) >> java.base@13.0.2 >> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) >> - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b >> - locked sun.nio.ch.EPollSelectorImpl@7ffc19 >> java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) >> app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) >> app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) >> app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) >> app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) >> app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >> app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) >> java.base@13.0.2/java.lang.Thread.run(Thread.java:830) >> >> >> This happens both when creating a connection directly or when using the >> plc4x connection pool. >> >> In both cases I am obtaining the connection, reading some modbus values, >> and then closing the connection. From the plcConnection metadata the >> connection type I am using is: >> >> org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection >> >> Over time thousands of these nioEventLoopGroup threads are created and I >> experience an OutOfMemory error in my program. >> >> Has anyone experienced this problem? Is there a way to force these threads >> to close, timeout, or to specify the total thread limit? >> >> Thanks and Regards, Adam >> >
Re: Leaking nioEventLoopGroup threads
Sure thing. Is 0.8.0-snapshot hosted anywhere or is that something that needs to be built? Regards Adam > On Aug 23, 2020, at 12:31 PM, Christofer Dutz > wrote: > > Hmm ... > > Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... 0.7.0 was > the first of the new generation drivers. We're maintaining the 0.6 branch and > working hard on making the new generation drivers 100% production ready. > > Chris > > > Am 23.08.20, 18:06 schrieb "Adam Rossi" : > >This is the latest 0.7.0 release from Maven. > > >Regards Adam > >> On Aug 23, 2020, at 11:56 AM, Christofer Dutz >> wrote: >> >> Hi Adam, >> >> which version of PLC4X are you using? I know we had similar reports some >> time ago, but had thought we had fixed them >> >> Chris >> >> >> >> Am 23.08.20, 16:40 schrieb "Adam Rossi" : >> >> Howdy. I am seeing a persistent thread being created for every >> plcConnection connect which looks like the following: >> >> Name: nioEventLoopGroup-11-1 >> State: RUNNABLE >> Total blocked: 0 Total waited: 0 >> >> Stack trace: >> java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) >> java.base@13.0.2 >> /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) >> java.base@13.0.2 >> /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) >> - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b >> - locked sun.nio.ch.EPollSelectorImpl@7ffc19 >> java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) >> >> app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) >> app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) >> app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) >> >> app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) >> >> app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) >> >> app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) >> java.base@13.0.2/java.lang.Thread.run(Thread.java:830) >> >> >> This happens both when creating a connection directly or when using the >> plc4x connection pool. >> >> In both cases I am obtaining the connection, reading some modbus values, >> and then closing the connection. From the plcConnection metadata the >> connection type I am using is: >> >> org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection >> >> Over time thousands of these nioEventLoopGroup threads are created and I >> experience an OutOfMemory error in my program. >> >> Has anyone experienced this problem? Is there a way to force these threads >> to close, timeout, or to specify the total thread limit? >> >> Thanks and Regards, Adam >> >
Re: Leaking nioEventLoopGroup threads
Hmm ... Could you possibly give 0.8.0-SNAPSHOT a try? or 0.6.x? ... 0.7.0 was the first of the new generation drivers. We're maintaining the 0.6 branch and working hard on making the new generation drivers 100% production ready. Chris Am 23.08.20, 18:06 schrieb "Adam Rossi" : This is the latest 0.7.0 release from Maven. Regards Adam > On Aug 23, 2020, at 11:56 AM, Christofer Dutz wrote: > > Hi Adam, > > which version of PLC4X are you using? I know we had similar reports some time ago, but had thought we had fixed them > > Chris > > > > Am 23.08.20, 16:40 schrieb "Adam Rossi" : > >Howdy. I am seeing a persistent thread being created for every >plcConnection connect which looks like the following: > >Name: nioEventLoopGroup-11-1 >State: RUNNABLE >Total blocked: 0 Total waited: 0 > >Stack trace: >java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) >java.base@13.0.2 >/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) >java.base@13.0.2 >/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) > - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b > - locked sun.nio.ch.EPollSelectorImpl@7ffc19 >java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) > app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) >app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) >app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) > app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) > app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) >java.base@13.0.2/java.lang.Thread.run(Thread.java:830) > > >This happens both when creating a connection directly or when using the >plc4x connection pool. > >In both cases I am obtaining the connection, reading some modbus values, >and then closing the connection. From the plcConnection metadata the >connection type I am using is: > >org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection > >Over time thousands of these nioEventLoopGroup threads are created and I >experience an OutOfMemory error in my program. > >Has anyone experienced this problem? Is there a way to force these threads >to close, timeout, or to specify the total thread limit? > >Thanks and Regards, Adam >
Re: Leaking nioEventLoopGroup threads
This is the latest 0.7.0 release from Maven. Regards Adam > On Aug 23, 2020, at 11:56 AM, Christofer Dutz > wrote: > > Hi Adam, > > which version of PLC4X are you using? I know we had similar reports some time > ago, but had thought we had fixed them > > Chris > > > > Am 23.08.20, 16:40 schrieb "Adam Rossi" : > >Howdy. I am seeing a persistent thread being created for every >plcConnection connect which looks like the following: > >Name: nioEventLoopGroup-11-1 >State: RUNNABLE >Total blocked: 0 Total waited: 0 > >Stack trace: >java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) >java.base@13.0.2 >/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) >java.base@13.0.2 >/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) > - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b > - locked sun.nio.ch.EPollSelectorImpl@7ffc19 >java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) > > app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) >app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) >app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) > > app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) > > app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) > > app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) >java.base@13.0.2/java.lang.Thread.run(Thread.java:830) > > >This happens both when creating a connection directly or when using the >plc4x connection pool. > >In both cases I am obtaining the connection, reading some modbus values, >and then closing the connection. From the plcConnection metadata the >connection type I am using is: > >org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection > >Over time thousands of these nioEventLoopGroup threads are created and I >experience an OutOfMemory error in my program. > >Has anyone experienced this problem? Is there a way to force these threads >to close, timeout, or to specify the total thread limit? > >Thanks and Regards, Adam >
Re: Leaking nioEventLoopGroup threads
Hi Adam, which version of PLC4X are you using? I know we had similar reports some time ago, but had thought we had fixed them Chris Am 23.08.20, 16:40 schrieb "Adam Rossi" : Howdy. I am seeing a persistent thread being created for every plcConnection connect which looks like the following: Name: nioEventLoopGroup-11-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b - locked sun.nio.ch.EPollSelectorImpl@7ffc19 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) This happens both when creating a connection directly or when using the plc4x connection pool. In both cases I am obtaining the connection, reading some modbus values, and then closing the connection. From the plcConnection metadata the connection type I am using is: org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection Over time thousands of these nioEventLoopGroup threads are created and I experience an OutOfMemory error in my program. Has anyone experienced this problem? Is there a way to force these threads to close, timeout, or to specify the total thread limit? Thanks and Regards, Adam
Leaking nioEventLoopGroup threads
Howdy. I am seeing a persistent thread being created for every plcConnection connect which looks like the following: Name: nioEventLoopGroup-11-1 State: RUNNABLE Total blocked: 0 Total waited: 0 Stack trace: java.base@13.0.2/sun.nio.ch.EPoll.wait(Native Method) java.base@13.0.2 /sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:120) java.base@13.0.2 /sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:124) - locked io.netty.channel.nio.SelectedSelectionKeySet@f6df6b - locked sun.nio.ch.EPollSelectorImpl@7ffc19 java.base@13.0.2/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:803) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@13.0.2/java.lang.Thread.run(Thread.java:830) This happens both when creating a connection directly or when using the plc4x connection pool. In both cases I am obtaining the connection, reading some modbus values, and then closing the connection. From the plcConnection metadata the connection type I am using is: org.apache.plc4x.java.spi.connection.DefaultNettyPlcConnection Over time thousands of these nioEventLoopGroup threads are created and I experience an OutOfMemory error in my program. Has anyone experienced this problem? Is there a way to force these threads to close, timeout, or to specify the total thread limit? Thanks and Regards, Adam