[ https://issues.apache.org/jira/browse/HADOOP-4386?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12638490#action_12638490 ]
Raghu Angadi commented on HADOOP-4386: -------------------------------------- Ideally no RPC should be blocked because of another 'zero copy RPC' (where RPC layer does not have control how fast the data is written or read). > RPC support for large data transfers. > ------------------------------------- > > Key: HADOOP-4386 > URL: https://issues.apache.org/jira/browse/HADOOP-4386 > Project: Hadoop Core > Issue Type: New Feature > Components: dfs, ipc > Reporter: Raghu Angadi > > Currently HDFS has a socket level protocol for serving HDFS data to clients. > Clients do not use RPCs to read or write data. Fundamentally there is no > reason why this data transfer can not use RPCs. > This jira is place holder for any porting Datanode transfers to RPC. This > topic has been discussed in varying detail many times, the latest being in > the context of HADOOP-3856. There are quite a few issues to be resolved both > at API level and at implementation level. > We should probably copy some of the comments from HADOOP-3856 to here. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.