[ https://issues.apache.org/jira/browse/HADOOP-1338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12649401#action_12649401 ]
Jothi Padmanabhan commented on HADOOP-1338: ------------------------------------------- bq. HTTP's keep-alive would keep the connection alive for only a certain time I believe I do not think there is a timeout for the duration for which the connection is kept alive -- Keep alive packets are sent out after a 'keepalive' time (normally 2 hours) of idleness and if the opposite side responds, the connection is kept alive. However, as pointed out, it might not be a good idea to just hold on to connections for scalability reasons. > Improve the shuffle phase by using the "connection: keep-alive" and doing > batch transfers of files > -------------------------------------------------------------------------------------------------- > > Key: HADOOP-1338 > URL: https://issues.apache.org/jira/browse/HADOOP-1338 > Project: Hadoop Core > Issue Type: Improvement > Components: mapred > Reporter: Devaraj Das > > We should do transfers of map outputs at the granularity of > *total-bytes-transferred* rather than the current way of transferring a > single file and then closing the connection to the server. A single > TaskTracker might have a couple of map output files for a given reduce, and > we should transfer multiple of them (upto a certain total size) in a single > connection to the TaskTracker. Using HTTP-1.1's keep-alive connection would > help since it would keep the connection open for more than one file transfer. > We should limit the transfers to a certain size so that we don't hold up a > jetty thread indefinitely (and cause timeouts for other clients). > Overall, this should give us improved performance. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.