[ 
https://issues.apache.org/jira/browse/HDFS-1457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12927259#action_12927259
 ] 

Lu Yilei commented on HDFS-1457:
--------------------------------

To Hairong. It is a good idea. If we transport  uncompression data, we should 
do it as same as your great idea. But If the fsimage is not compressed, It will 
be better for limit transmission.
We use the method of limit and compression transmission in Baidu. And it runs 
well. 

> Limit transmission rate when transfering image between primary and secondary 
> NNs
> --------------------------------------------------------------------------------
>
>                 Key: HDFS-1457
>                 URL: https://issues.apache.org/jira/browse/HDFS-1457
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: name-node
>    Affects Versions: 0.22.0
>            Reporter: Hairong Kuang
>             Fix For: 0.22.0
>
>         Attachments: checkpoint-limitandcompress.patch, 
> trunkThrottleImage.patch
>
>
> If the fsimage is very big. The network is full in a short time when 
> SeconaryNamenode do checkpoint, leading to Jobtracker access Namenode to get 
> relevant file data to fail in job initialization phase. So we limit 
> transmission speed and compress transmission to resolve the problem. 

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to