[
https://issues.apache.org/jira/browse/HADOOP-5825?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12710275#action_12710275
]
dhruba borthakur commented on HADOOP-5825:
------------------------------------------
Is it possible that FSNamesystem.removePathAndBlocks() is the major bottleneck?
If so, we can maybe re-arrange the code to keep the "freeing up resource" part
of the code outside the FSNamesystem lock.
> Recursively deleting a directory with millions of files makes NameNode
> unresponsive for other commands until the deletion completes
> -----------------------------------------------------------------------------------------------------------------------------------
>
> Key: HADOOP-5825
> URL: https://issues.apache.org/jira/browse/HADOOP-5825
> Project: Hadoop Core
> Issue Type: Bug
> Components: dfs
> Reporter: Suresh Srinivas
> Assignee: Suresh Srinivas
>
> Delete a directory with millions of files. This could take several minutes
> (observed 12 mins for 9 million files). While the operation is in progress
> FSNamesystem lock is held and the requests from clients are not handled until
> deletion completes.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.