Re: File leak in 7.0.23?

2012-01-02 Thread Mike Wertheim
If a simple test case isn't discovered, I'm happy to test out
potential fixes (by deploying the fixes to one of my production
servers and seeing whether or not the server dies after running for a
day).

On Sun, Jan 1, 2012 at 7:17 AM,  ma...@apache.org wrote:
 Janne Jalkanen janne.jalka...@ecyrd.com wrote:

APR + native. Good catch there, I took apr out and I am no longer
seeing the FD leak.

 OK. Sounds like APR/native has an issue. There was a fair bit of refactoring 
 in 7.0.22.

 I'll see if I can reproduce it. A simple test case may help.

 Mark



 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-02 Thread Christopher Schultz
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Janne,

On 1/1/12 12:23 PM, Janne Jalkanen wrote:
 OK. Sounds like APR/native has an issue. There was a fair bit of
 refactoring in 7.0.22.
 
 I'll see if I can reproduce it. A simple test case may help.
 
 Should I open a bug now on the issue tracker?

Yes. That would be the best place to post a simple test case (in the
form of a webapp, please).

- -chris
-BEGIN PGP SIGNATURE-
Version: GnuPG/MacGPG2 v2.0.17 (Darwin)
Comment: GPGTools - http://gpgtools.org
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/

iEYEARECAAYFAk8CPAYACgkQ9CaO5/Lv0PDp4gCfUHRy2J1xzC4w0rxzsfaWRMmm
WJEAn0huK8QK5Bx/93QRXhhT0XI0goyA
=fHq4
-END PGP SIGNATURE-

-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-02 Thread Konstantin Kolinko
2011/12/31 Janne Jalkanen janne.jalka...@ecyrd.com:
 Hi all!

 I am seeing odd behaviour with 7.0.23, with the tomcat user's open file count 
 increasing slowly, but consistently. Two other instances running the exact 
 same codebase on identical hardware, BUT with Tomcat 7.0.20, are not 
 exhibiting the same behaviour. 7.0.20 is rock solid, 7.0.23 dies due to too 
 many open files every now and then. I've increased ulimit for now, but this 
 is still a bit nasty.

 Any suggestions where to look? Should I file a bug?

 lsof says

 ...
 java    21299 ubuntu   87r     sock                0,6       0t0  31983006 
 can't identify protocol
 java    21299 ubuntu   88r     sock                0,6       0t0  31983007 
 can't identify protocol
 java    21299 ubuntu   89r     sock                0,6       0t0  31983008 
 can't identify protocol
 java    21299 ubuntu   90r     sock                0,6       0t0  31989046 
 can't identify protocol
 java    21299 ubuntu   91r     sock                0,6       0t0  31986504 
 can't identify protocol
 java    21299 ubuntu   92r     sock                0,6       0t0  31987223 
 can't identify protocol
 ...

 with a new one every couple of minutes.

I wonder whether it is possible to get some numbers in Access Log to
match with lsof output. So to get a clue what kinds of requests result
in the leak.

a) normal dynamic requests
b) comet or asynchronous requests
c) requests for static files that are served via sendfile

Maybe there are timeouts, or connections closed at client's side.


It should be easy to exclude sendfile:

a) When sendfile is used it is visible in access log, because a static
file is requested and file size that is logged as 0 (until 7.0.24).
See https://issues.apache.org/bugzilla/show_bug.cgi?id=52316

b) If it easy to turn it off, by setting useSendfile=false on a Connector.


 No OOM issues, no errors in log files, until the eventual SocketException 
 when ulimit is reached:

I wonder whether there were some thread deaths. If that happens the
ThreadGroup.uncaughtException() method will print the stacktrace
directly into System.err. That is it will be written to catalina.out
file (and it wouldn't be written to catalina.log).

Best regards,
Konstantin Kolinko

-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-02 Thread Mike Wertheim
On Mon, Jan 2, 2012 at 4:31 PM, Konstantin Kolinko
knst.koli...@gmail.com wrote:
 2011/12/31 Janne Jalkanen janne.jalka...@ecyrd.com:
 Hi all!

 I am seeing odd behaviour with 7.0.23, with the tomcat user's open file 
 count increasing slowly, but consistently. Two other instances running the 
 exact same codebase on identical hardware, BUT with Tomcat 7.0.20, are not 
 exhibiting the same behaviour. 7.0.20 is rock solid, 7.0.23 dies due to too 
 many open files every now and then. I've increased ulimit for now, but this 
 is still a bit nasty.

 Any suggestions where to look? Should I file a bug?

 lsof says

 ...
 java    21299 ubuntu   87r     sock                0,6       0t0  31983006 
 can't identify protocol
 java    21299 ubuntu   88r     sock                0,6       0t0  31983007 
 can't identify protocol
 java    21299 ubuntu   89r     sock                0,6       0t0  31983008 
 can't identify protocol
 java    21299 ubuntu   90r     sock                0,6       0t0  31989046 
 can't identify protocol
 java    21299 ubuntu   91r     sock                0,6       0t0  31986504 
 can't identify protocol
 java    21299 ubuntu   92r     sock                0,6       0t0  31987223 
 can't identify protocol
 ...

 with a new one every couple of minutes.

 I wonder whether it is possible to get some numbers in Access Log to
 match with lsof output. So to get a clue what kinds of requests result
 in the leak.

 a) normal dynamic requests
 b) comet or asynchronous requests
 c) requests for static files that are served via sendfile

For my app (which fails on 7.0.23 and uses the APR Connector), there
are no comet requests.  99% of the requests are normal dynamic
requests, and the remaining 1% are static files.



 Maybe there are timeouts, or connections closed at client's side.


 It should be easy to exclude sendfile:

 a) When sendfile is used it is visible in access log, because a static
 file is requested and file size that is logged as 0 (until 7.0.24).
 See https://issues.apache.org/bugzilla/show_bug.cgi?id=52316

 b) If it easy to turn it off, by setting useSendfile=false on a Connector.


 No OOM issues, no errors in log files, until the eventual SocketException 
 when ulimit is reached:

 I wonder whether there were some thread deaths. If that happens the
 ThreadGroup.uncaughtException() method will print the stacktrace
 directly into System.err. That is it will be written to catalina.out
 file (and it wouldn't be written to catalina.log).

 Best regards,
 Konstantin Kolinko

 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-01 Thread Janne Jalkanen

APR + native. Good catch there, I took apr out and I am no longer seeing the FD 
leak. 

I was running APR 1.3.8 with TC-native 1.1.19, so I thought that maybe these 
old versions were the problem. But I upgraded to APR 1.4.5 and tc-native 1.1.22 
(the latest stables from both) and am still seeing the FD leak, whenever 
APR+native is enabled.

/Janne

On Dec 31, 2011, at 23:05 , Mark Thomas wrote:

 On 31/12/2011 13:38, Janne Jalkanen wrote:
 Which Connector are you using?
 
Connector port=8080 protocol=HTTP/1.1 
   connectionTimeout=2 
   redirectPort=8443 
   URIEncoding=UTF-8
   compression=on
   
 compressableMimeType=text/html,text/plain,text/css,text/javascript,application/json,application/javascript
   noCompressionUserAgents=.*MSIE 6.*/
 
 Is that APR/native or BIO?
 
 Mark
 
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-01 Thread markt
Janne Jalkanen janne.jalka...@ecyrd.com wrote:

APR + native. Good catch there, I took apr out and I am no longer
seeing the FD leak. 

OK. Sounds like APR/native has an issue. There was a fair bit of refactoring in 
7.0.22.

I'll see if I can reproduce it. A simple test case may help.

Mark



-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-01 Thread Janne Jalkanen
 If you take heap dumps can you see an Increase in the number of sockets?

What should I be looking for, assuming that the problem is with APR connection? 
Any particular class?

/Janne
-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2012-01-01 Thread Janne Jalkanen
 OK. Sounds like APR/native has an issue. There was a fair bit of refactoring 
 in 7.0.22.
 
 I'll see if I can reproduce it. A simple test case may help.

Should I open a bug now on the issue tracker?

/Janne
-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



File leak in 7.0.23?

2011-12-31 Thread Janne Jalkanen
Hi all!

I am seeing odd behaviour with 7.0.23, with the tomcat user's open file count 
increasing slowly, but consistently. Two other instances running the exact same 
codebase on identical hardware, BUT with Tomcat 7.0.20, are not exhibiting the 
same behaviour. 7.0.20 is rock solid, 7.0.23 dies due to too many open files 
every now and then. I've increased ulimit for now, but this is still a bit 
nasty.

Any suggestions where to look? Should I file a bug?

lsof says

...
java21299 ubuntu   87r sock0,6   0t0  31983006 
can't identify protocol
java21299 ubuntu   88r sock0,6   0t0  31983007 
can't identify protocol
java21299 ubuntu   89r sock0,6   0t0  31983008 
can't identify protocol
java21299 ubuntu   90r sock0,6   0t0  31989046 
can't identify protocol
java21299 ubuntu   91r sock0,6   0t0  31986504 
can't identify protocol
java21299 ubuntu   92r sock0,6   0t0  31987223 
can't identify protocol
...

with a new one every couple of minutes.

java version 1.6.0_26
Java(TM) SE Runtime Environment (build 1.6.0_26-b03)
Java HotSpot(TM) 64-Bit Server VM (build 20.1-b02, mixed mode)

/usr/bin/java -Dnop -server -Xmx1024m -Xms128m -XX:MaxPermSize=256m 
-Dcom.sun.management.jmxremote.port=redacted 
-Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager 
-Djava.endorsed.dirs=/home/ubuntu/tomcat-bin/current/endorsed -classpath 
/home/ubuntu/tomcat-bin/current/bin/bootstrap.jar:/home/ubuntu/tomcat-bin/current/bin/tomcat-juli.jar
 -Dcatalina.base=/home/ubuntu/tomcat-run 
-Dcatalina.home=/home/ubuntu/tomcat-bin/current 
-Djava.io.tmpdir=/home/ubuntu/tomcat-run/temp 
org.apache.catalina.startup.Bootstrap start

OS: Ubuntu 10.04 LTS, kernel 2.6.32 smp.

No OOM issues, no errors in log files, until the eventual SocketException when 
ulimit is reached:

30-Dec-2011 19:05:47 sun.rmi.transport.tcp.TCPTransport$AcceptLoop 
executeAcceptLoop
WARNING: RMI TCP Accept-0: accept loop for 
ServerSocket[addr=0.0.0.0/0.0.0.0,port=0,localport=41709] throws
java.net.SocketException: Too many open files
at java.net.PlainSocketImpl.socketAccept(Native Method)
at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408)
at java.net.ServerSocket.implAccept(ServerSocket.java:462)
at java.net.ServerSocket.accept(ServerSocket.java:430)
at 
sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:369)

at 
sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:341)
at java.lang.Thread.run(Thread.java:662)

/Janne
-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Pid
On 31/12/2011 13:06, Janne Jalkanen wrote:
 Hi all!
 
 I am seeing odd behaviour with 7.0.23, with the tomcat user's open file count 
 increasing slowly, but consistently. Two other instances running the exact 
 same codebase on identical hardware, BUT with Tomcat 7.0.20, are not 
 exhibiting the same behaviour. 7.0.20 is rock solid, 7.0.23 dies due to too 
 many open files every now and then. I've increased ulimit for now, but this 
 is still a bit nasty.
 
 Any suggestions where to look? Should I file a bug?

Which Connector are you using?

What is the minimum that is required to reproduce the issue?


 lsof says

Did you supply any options to lsof?

 
 ...
 java21299 ubuntu   87r sock0,6   0t0  31983006 
 can't identify protocol
 java21299 ubuntu   88r sock0,6   0t0  31983007 
 can't identify protocol
 java21299 ubuntu   89r sock0,6   0t0  31983008 
 can't identify protocol
 java21299 ubuntu   90r sock0,6   0t0  31989046 
 can't identify protocol
 java21299 ubuntu   91r sock0,6   0t0  31986504 
 can't identify protocol
 java21299 ubuntu   92r sock0,6   0t0  31987223 
 can't identify protocol
 ...
 
 with a new one every couple of minutes.
 
 java version 1.6.0_26
 Java(TM) SE Runtime Environment (build 1.6.0_26-b03)
 Java HotSpot(TM) 64-Bit Server VM (build 20.1-b02, mixed mode)
 
 /usr/bin/java -Dnop -server -Xmx1024m -Xms128m -XX:MaxPermSize=256m 
 -Dcom.sun.management.jmxremote.port=redacted 
 -Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager 
 -Djava.endorsed.dirs=/home/ubuntu/tomcat-bin/current/endorsed -classpath 
 /home/ubuntu/tomcat-bin/current/bin/bootstrap.jar:/home/ubuntu/tomcat-bin/current/bin/tomcat-juli.jar
  -Dcatalina.base=/home/ubuntu/tomcat-run 
 -Dcatalina.home=/home/ubuntu/tomcat-bin/current 
 -Djava.io.tmpdir=/home/ubuntu/tomcat-run/temp 
 org.apache.catalina.startup.Bootstrap start
 
 OS: Ubuntu 10.04 LTS, kernel 2.6.32 smp.
 
 No OOM issues, no errors in log files, until the eventual SocketException 
 when ulimit is reached:
 
 30-Dec-2011 19:05:47 sun.rmi.transport.tcp.TCPTransport$AcceptLoop 
 executeAcceptLoop
 WARNING: RMI TCP Accept-0: accept loop for 
 ServerSocket[addr=0.0.0.0/0.0.0.0,port=0,localport=41709] throws

Are you using RMI for anything in your application or is this a
reference to the JMX connection?

Are you using JMX for something specific?


p

 java.net.SocketException: Too many open files
 at java.net.PlainSocketImpl.socketAccept(Native Method)
 at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:408)
 at java.net.ServerSocket.implAccept(ServerSocket.java:462)
 at java.net.ServerSocket.accept(ServerSocket.java:430)
 at 
 sun.rmi.transport.tcp.TCPTransport$AcceptLoop.executeAcceptLoop(TCPTransport.java:369)
 
 at 
 sun.rmi.transport.tcp.TCPTransport$AcceptLoop.run(TCPTransport.java:341)
 at java.lang.Thread.run(Thread.java:662)
 
 /Janne
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org
 


-- 

[key:62590808]



signature.asc
Description: OpenPGP digital signature


Re: File leak in 7.0.23?

2011-12-31 Thread Janne Jalkanen
 Which Connector are you using?

Connector port=8080 protocol=HTTP/1.1 
   connectionTimeout=2 
   redirectPort=8443 
   URIEncoding=UTF-8
   compression=on
   
compressableMimeType=text/html,text/plain,text/css,text/javascript,application/json,application/javascript
   noCompressionUserAgents=.*MSIE 6.*/


 What is the minimum that is required to reproduce the issue?

Wish I knew. This is from our production cluster, so I'm a bit hesitant to 
start experimenting with it. However, there was no configuration change between 
7.0.20 and 7.0.23, I just shut down tomcat, unpacked the 7.0.23 tar.gz, changed 
symlink to point at the new binaries, and restarted.

 
 
 lsof says
 
 Did you supply any options to lsof?

lsof -nP -u ubuntu|grep 21299|grep identify protocol

 Are you using RMI for anything in your application or is this a
 reference to the JMX connection?

I am using RMI only for JMX monitoring. The socket exception times *do* roughly 
correspond to Munin contacting my application every five minutes for getting 
monitoring data, which is logical because they would be requiring a set of new 
file handles.

Ah yes, I also have Terracotta session clustering configured using their Valve. 
That might be using RMI, but I would be expecting to see a lot more errors in 
that case.

/Janne


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Pid
On 31/12/2011 13:38, Janne Jalkanen wrote:
 Which Connector are you using?
 
 Connector port=8080 protocol=HTTP/1.1 
connectionTimeout=2 
redirectPort=8443 
URIEncoding=UTF-8
compression=on

 compressableMimeType=text/html,text/plain,text/css,text/javascript,application/json,application/javascript
noCompressionUserAgents=.*MSIE 6.*/
 
 
 What is the minimum that is required to reproduce the issue?
 
 Wish I knew. This is from our production cluster, so I'm a bit hesitant to 
 start experimenting with it. However, there was no configuration change 
 between 7.0.20 and 7.0.23, I just shut down tomcat, unpacked the 7.0.23 
 tar.gz, changed symlink to point at the new binaries, and restarted.
 


 lsof says

 Did you supply any options to lsof?
 
 lsof -nP -u ubuntu|grep 21299|grep identify protocol
 
 Are you using RMI for anything in your application or is this a
 reference to the JMX connection?
 
 I am using RMI only for JMX monitoring. The socket exception times *do* 
 roughly correspond to Munin contacting my application every five minutes for 
 getting monitoring data, which is logical because they would be requiring a 
 set of new file handles.

When did the problem start occurring and what else has changed?

Can you stop monitoring with Munin and see if the problem goes away?
If it does, I would consider that Munin may not be properly closing the
connections it makes.


p

 Ah yes, I also have Terracotta session clustering configured using their 
 Valve. That might be using RMI, but I would be expecting to see a lot more 
 errors in that case.
 
 /Janne
 
 
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org
 


-- 

[key:62590808]



signature.asc
Description: OpenPGP digital signature


Re: File leak in 7.0.23?

2011-12-31 Thread Janne Jalkanen
 When did the problem start occurring and what else has changed?

Exactly at the time when I upgraded to 7.0.23.  I don't recall making any other 
modifications (I would've suspected them first ;-)

I will try to downgrade to 7.0.22 and lower to try and see if there's a 
difference between Tomcat versions.

 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.

I did, and the problem did not go away. Open file count still increasing.

/Janne


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread André Warnier

Janne Jalkanen wrote:

When did the problem start occurring and what else has changed?


Exactly at the time when I upgraded to 7.0.23.  I don't recall making any other 
modifications (I would've suspected them first ;-)

I will try to downgrade to 7.0.22 and lower to try and see if there's a 
difference between Tomcat versions.


Can you stop monitoring with Munin and see if the problem goes away?
If it does, I would consider that Munin may not be properly closing the
connections it makes.


I did, and the problem did not go away. Open file count still increasing.


What if you remove the command-line switch
-Dcom.sun.management.jmxremote.port=redacted
?

(just trying to figure out what these sockets are..)

Also, what does
netstat -pan | grep tomcat pid
have to say ?

-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Mike Wertheim
I'm not sure how useful this comment is, but...  I also recently
posted about an app that runs fine on Tomcat 7.0.21 and dies a slow
horrible death on Tomcat 7.0.23.  It would seem that a bug was
introduced in either 7.0.22 or 7.0.23.



On Sat, Dec 31, 2011 at 6:48 AM, André Warnier a...@ice-sa.com wrote:
 Janne Jalkanen wrote:

 When did the problem start occurring and what else has changed?


 Exactly at the time when I upgraded to 7.0.23.  I don't recall making any
 other modifications (I would've suspected them first ;-)

 I will try to downgrade to 7.0.22 and lower to try and see if there's a
 difference between Tomcat versions.

 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.


 I did, and the problem did not go away. Open file count still increasing.

 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

 (just trying to figure out what these sockets are..)

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Janne Jalkanen

ATM my testing looks like 7.0.21 works, whereas 7.0.22 is leaky. Will continue 
to investigate.

/Janne

On Dec 31, 2011, at 20:18 , Mike Wertheim wrote:

 I'm not sure how useful this comment is, but...  I also recently
 posted about an app that runs fine on Tomcat 7.0.21 and dies a slow
 horrible death on Tomcat 7.0.23.  It would seem that a bug was
 introduced in either 7.0.22 or 7.0.23.
 
 
 
 On Sat, Dec 31, 2011 at 6:48 AM, André Warnier a...@ice-sa.com wrote:
 Janne Jalkanen wrote:
 
 When did the problem start occurring and what else has changed?
 
 
 Exactly at the time when I upgraded to 7.0.23.  I don't recall making any
 other modifications (I would've suspected them first ;-)
 
 I will try to downgrade to 7.0.22 and lower to try and see if there's a
 difference between Tomcat versions.
 
 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.
 
 
 I did, and the problem did not go away. Open file count still increasing.
 
 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?
 
 (just trying to figure out what these sockets are..)
 
 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?
 
 
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org
 
 
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Pid *
On 31 Dec 2011, at 18:19, Mike Wertheim m...@hyperreal.org wrote:

 I'm not sure how useful this comment is, but...  I also recently
 posted about an app that runs fine on Tomcat 7.0.21 and dies a slow
 horrible death on Tomcat 7.0.23.  It would seem that a bug was
 introduced in either 7.0.22 or 7.0.23.

There are now 3 threads with non-obvious issues referencing 7.0.23. I
don't yet see any common ground.

Following them all and comparing notes isn't unreasonable.


p


 On Sat, Dec 31, 2011 at 6:48 AM, André Warnier a...@ice-sa.com wrote:
 Janne Jalkanen wrote:

 When did the problem start occurring and what else has changed?


 Exactly at the time when I upgraded to 7.0.23.  I don't recall making any
 other modifications (I would've suspected them first ;-)

 I will try to downgrade to 7.0.22 and lower to try and see if there's a
 difference between Tomcat versions.

 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.


 I did, and the problem did not go away. Open file count still increasing.

 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

 (just trying to figure out what these sockets are..)

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Mike Wertheim
Janne's latest email says that 7.0.22 is leaky as well.  So the
regression most likely happened between 7.0.21 and 7.0.22.

I'm not familiar with the Tomcat code base.  But I wonder how
difficult it would be for someone to review all of the code changes
that were checked in between 7.0.21 and 7.0.22, keeping an eye out for
anything that could be causing a file leak.



On Sat, Dec 31, 2011 at 11:17 AM, Pid * p...@pidster.com wrote:
 On 31 Dec 2011, at 18:19, Mike Wertheim m...@hyperreal.org wrote:

 I'm not sure how useful this comment is, but...  I also recently
 posted about an app that runs fine on Tomcat 7.0.21 and dies a slow
 horrible death on Tomcat 7.0.23.  It would seem that a bug was
 introduced in either 7.0.22 or 7.0.23.

 There are now 3 threads with non-obvious issues referencing 7.0.23. I
 don't yet see any common ground.

 Following them all and comparing notes isn't unreasonable.


 p


 On Sat, Dec 31, 2011 at 6:48 AM, André Warnier a...@ice-sa.com wrote:
 Janne Jalkanen wrote:

 When did the problem start occurring and what else has changed?


 Exactly at the time when I upgraded to 7.0.23.  I don't recall making any
 other modifications (I would've suspected them first ;-)

 I will try to downgrade to 7.0.22 and lower to try and see if there's a
 difference between Tomcat versions.

 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.


 I did, and the problem did not go away. Open file count still increasing.

 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

 (just trying to figure out what these sockets are..)

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Janne Jalkanen
 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

No effect.

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?

Nothing unusual (i.e. stuff I wouldn't expect) or different from the other 
instances: Database connections, listening on 8080, etc.

It looks like 7.0.2021 have constantly a couple of FDs open that are in the 
can't identify protocol -domain; with .22 the number increases slowly, and 
with .23 it increases a bit more rapidly.  The rate of change apparently seems 
to correlate to the number of requests served by Tomcat.

/Janne
-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Pid *
On 31 Dec 2011, at 20:03, Mike Wertheim m...@hyperreal.org wrote:

 Janne's latest email says that 7.0.22 is leaky as well.  So the
 regression most likely happened between 7.0.21 and 7.0.22.

 I'm not familiar with the Tomcat code base.  But I wonder how
 difficult it would be for someone to review all of the code changes
 that were checked in between 7.0.21 and 7.0.22, keeping an eye out for
 anything that could be causing a file leak.

Right now  for me: hard, because I'm using an iPhone.

Tomorrow easier: but having a minimal test case for reproducing the
error will make things easier.


p


(p.s. please don't top-post)

 On Sat, Dec 31, 2011 at 11:17 AM, Pid * p...@pidster.com wrote:
 On 31 Dec 2011, at 18:19, Mike Wertheim m...@hyperreal.org wrote:

 I'm not sure how useful this comment is, but...  I also recently
 posted about an app that runs fine on Tomcat 7.0.21 and dies a slow
 horrible death on Tomcat 7.0.23.  It would seem that a bug was
 introduced in either 7.0.22 or 7.0.23.

 There are now 3 threads with non-obvious issues referencing 7.0.23. I
 don't yet see any common ground.

 Following them all and comparing notes isn't unreasonable.


 p


 On Sat, Dec 31, 2011 at 6:48 AM, André Warnier a...@ice-sa.com wrote:
 Janne Jalkanen wrote:

 When did the problem start occurring and what else has changed?


 Exactly at the time when I upgraded to 7.0.23.  I don't recall making any
 other modifications (I would've suspected them first ;-)

 I will try to downgrade to 7.0.22 and lower to try and see if there's a
 difference between Tomcat versions.

 Can you stop monitoring with Munin and see if the problem goes away?
 If it does, I would consider that Munin may not be properly closing the
 connections it makes.


 I did, and the problem did not go away. Open file count still increasing.

 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

 (just trying to figure out what these sockets are..)

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Pid *
On 31 Dec 2011, at 20:05, Janne Jalkanen janne.jalka...@ecyrd.com wrote:

 What if you remove the command-line switch
 -Dcom.sun.management.jmxremote.port=redacted
 ?

 No effect.

 Also, what does
 netstat -pan | grep tomcat pid
 have to say ?

 Nothing unusual (i.e. stuff I wouldn't expect) or different from the other 
 instances: Database connections, listening on 8080, etc.

 It looks like 7.0.2021 have constantly a couple of FDs open that are in the 
 can't identify protocol -domain; with .22 the number increases slowly, and 
 with .23 it increases a bit more rapidly.  The rate of change apparently 
 seems to correlate to the number of requests served by Tomcat.

If you take heap dumps can you see an Increase in the number of sockets?


p


 /Janne
 -
 To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
 For additional commands, e-mail: users-h...@tomcat.apache.org


-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org



Re: File leak in 7.0.23?

2011-12-31 Thread Mark Thomas
On 31/12/2011 13:38, Janne Jalkanen wrote:
 Which Connector are you using?
 
 Connector port=8080 protocol=HTTP/1.1 
connectionTimeout=2 
redirectPort=8443 
URIEncoding=UTF-8
compression=on

 compressableMimeType=text/html,text/plain,text/css,text/javascript,application/json,application/javascript
noCompressionUserAgents=.*MSIE 6.*/

Is that APR/native or BIO?

Mark

-
To unsubscribe, e-mail: users-unsubscr...@tomcat.apache.org
For additional commands, e-mail: users-h...@tomcat.apache.org