[ https://issues.apache.org/jira/browse/HADOOP-1849?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12526517 ]
Raghu Angadi commented on HADOOP-1849: -------------------------------------- At least while testing, if this is configurable, it would be easy to asks users to experiment with different values. > IPC server max queue size should be configurable > ------------------------------------------------ > > Key: HADOOP-1849 > URL: https://issues.apache.org/jira/browse/HADOOP-1849 > Project: Hadoop > Issue Type: Improvement > Components: ipc > Reporter: Raghu Angadi > > Currently max queue size for IPC server is set to (100 * handlers). Usually > when RPC failures are observed (e.g. HADOOP-1763), we increase number of > handlers and the problem goes away. I think a big part of such a fix is > increase in max queue size. I think we should make maxQsize per handler > configurable (with a bigger default than 100). There are other improvements > also (HADOOP-1841). > Server keeps reading RPC requests from clients. When the number in-flight > RPCs is larger than maxQsize, the earliest RPCs are deleted. This is the main > feedback Server has for the client. I have often heard from users that Hadoop > doesn't handle bursty traffic. > Say handler count is 10 (default) and Server can handle 1000 RPCs a sec > (quite conservative/low for a typical server), it implies that an RPC can > wait for only for 1 sec before it is dropped. If there 3000 clients and all > of them send RPCs around the same time (not very rare, with heartbeats etc), > 2000 will be dropped. In stead of dropping the earliest RPCs, if the server > delays reading new RPCs, the feedback to clients would be much smoother, I > will file another jira regd queue management. > For this jira I propose to make queue size per handler configurable, with a > larger default (may be 500). -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.