[ 
https://issues.apache.org/jira/browse/HADOOP-771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12497009
 ] 

Milind Bhandarkar commented on HADOOP-771:
------------------------------------------

>> On the other hand, won't an RPC per file cause a lot of namenode traffic?

yes, it would. But only when user intentionally deletes all his/her data, which 
I believe happens rarely.

I think, DFS is optimized for a small number of large files, rather than large 
number of small files, so in practice, this erasure of entire tree should not 
happen frequently.

> Namenode should return error when trying to delete non-empty directory
> ----------------------------------------------------------------------
>
>                 Key: HADOOP-771
>                 URL: https://issues.apache.org/jira/browse/HADOOP-771
>             Project: Hadoop
>          Issue Type: Improvement
>          Components: dfs
>    Affects Versions: 0.8.0
>         Environment: all
>            Reporter: Milind Bhandarkar
>         Assigned To: Sameer Paranjpye
>
> Currently, the namenode.delete() method allows recursive deletion of a 
> directory. That is, even a non-empty directory could be deleted using 
> namenode.delete(). To avoid costly programmer errors, the namenode should not 
> remove the non-empty directories in this method. Recursively deleting 
> directory should either be performed with listPaths() followed by a delete() 
> for every path, or with a specific namenode method such as deleteRecursive().

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to