[ 
https://issues.apache.org/jira/browse/HDFS-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14165849#comment-14165849
 ] 

Colin Patrick McCabe commented on HDFS-6919:
--------------------------------------------

I'd really like to see us shrink the write cache size dynamically based on 
things entering the read cache, as we discussed in the last few comments.

Since this could be considered an incompatible change, I'd really really like 
to see it happen before this goes into 2.6.  I don't think it's that much work. 
 If the schedule makes this impossible, then we can at least add a release note 
that this behavior will be implemented soon, so that we can do it in 2.6.1 (or 
whatever follow-on release) without users being surprised.

> Enforce a single limit for RAM disk usage and replicas cached via locking
> -------------------------------------------------------------------------
>
>                 Key: HDFS-6919
>                 URL: https://issues.apache.org/jira/browse/HDFS-6919
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>            Reporter: Arpit Agarwal
>            Assignee: Colin Patrick McCabe
>            Priority: Blocker
>
> The DataNode can have a single limit for memory usage which applies to both 
> replicas cached via CCM and replicas on RAM disk.
> See comments 
> [1|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106025&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106025],
>  
> [2|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106245&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106245]
>  and 
> [3|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106575&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106575]
>  for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to