[ 
https://issues.apache.org/jira/browse/HDFS-6919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14152218#comment-14152218
 ] 

Colin Patrick McCabe commented on HDFS-6919:
--------------------------------------------

I like the idea of using the cache pool concept to control how much space we 
should use for HDFS-6581.  One reason is that cache pools can be modified at 
runtime through {{hdfs cacheadmin}}, whereas changing configuration keys 
normally requires a restart.  Another reason is because HDFS-4949 is managed 
via cache pools, it just makes more sense to use them for HDFS-6581 as well.

I have to think about what the best way to do this is.  One obvious way to go 
is to make a cache pool named "lazyPersist" and have this determine how much 
space we use for HDFS-6581.

> Enforce a single limit for RAM disk usage and replicas cached via locking
> -------------------------------------------------------------------------
>
>                 Key: HDFS-6919
>                 URL: https://issues.apache.org/jira/browse/HDFS-6919
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>            Reporter: Arpit Agarwal
>            Assignee: Colin Patrick McCabe
>            Priority: Blocker
>
> The DataNode can have a single limit for memory usage which applies to both 
> replicas cached via CCM and replicas on RAM disk.
> See comments 
> [1|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106025&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106025],
>  
> [2|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106245&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106245]
>  and 
> [3|https://issues.apache.org/jira/browse/HDFS-6581?focusedCommentId=14106575&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14106575]
>  for discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to