[ 
https://issues.apache.org/jira/browse/HADOOP-3328?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Raghu Angadi updated HADOOP-3328:
---------------------------------

    Attachment: HADOOP-3328.patch

Updated patch for trunk.

> DFS write pipeline : only the last datanode needs to verify checksum
> --------------------------------------------------------------------
>
>                 Key: HADOOP-3328
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3328
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: dfs
>    Affects Versions: 0.16.0
>            Reporter: Raghu Angadi
>            Assignee: Raghu Angadi
>         Attachments: HADOOP-3328.patch, HADOOP-3328.patch
>
>
> Currently all the datanodes in DFS write pipeline verify checksum. Since the 
> current protocol includes acks from  the datanodes, an ack from the last node 
> could also serve as verification that checksum ok. In that sense, only the 
> last datanode needs to verify checksum. Based on [this 
> comment|http://issues.apache.org/jira/browse/HADOOP-1702?focusedCommentId=12575553#action_12575553]
>  from HADOOP-1702, CPU consumption might go down by another 25-30% (4/14) 
> after HADOOP-1702. 
> Also this would make it easier to use transferTo() and transferFrom() on 
> intermediate datanodes since they don't need to look at the data.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to