[ 
https://issues.apache.org/jira/browse/HDFS-5589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13861804#comment-13861804
 ] 

Andrew Wang commented on HDFS-5589:
-----------------------------------

When we're populating {{possibilities}}, we check the DNs for validity, 
including having enough remaining capacity, so I think this is technically 
right. I agree though that it reads poorly, so I'll refactor this, and also add 
a test that tries to cache some big files.

> Namenode loops caching and uncaching when data should be uncached
> -----------------------------------------------------------------
>
>                 Key: HDFS-5589
>                 URL: https://issues.apache.org/jira/browse/HDFS-5589
>             Project: Hadoop HDFS
>          Issue Type: Sub-task
>          Components: caching, namenode
>    Affects Versions: 3.0.0
>            Reporter: Andrew Wang
>            Assignee: Andrew Wang
>         Attachments: hdfs-5589-1.patch
>
>
> This was reported by [~cnauroth] and [~brandonli], and [~schu] repro'd it too.
> If you add a new caching directive then remove it, the Namenode will 
> sometimes get stuck in a loop where it sends DNA_CACHE and then DNA_UNCACHE 
> repeatedly to the datanodes where the data was previously cached.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)

Reply via email to