Github user revans2 commented on the pull request:
https://github.com/apache/storm/pull/311#issuecomment-63495857
@ptgoetz I traced this down in the netty code to the second parameter to
the call to bind in the socket.
https://docs.oracle.com/javase/7/docs/api/java/net/ServerSocket.html#bind%28java.net.SocketAddress,%20int%29
It sets the maximum incoming connection queue. That way if the boss thread
is unable to keep up with accepting new connections the OS will keep them
buffered for a while until it can get to them. This should only be an issue
when lots of connections are being established very quickly, which would only
happen for very large topologies.
@caofangkun it would be good to add a better explanation for the config to
the documentation for the config. I would also like it if we could rename the
config to something like ```storm.messaging.netty.socket.backlog``` as I can
see another backlog being created in the future for tuples, instead of TCP
connection requests.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---