[ https://issues.apache.org/jira/browse/HADOOP-8973?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592830#comment-13592830 ]
Chris Nauroth commented on HADOOP-8973: --------------------------------------- That's interesting. Yes, I believe this is a bug in the existing code for the other overload of {{DiskChecker#checkDir}}. For example, suppose a dfs.datanode.data.dir on the local file system with owner "foo" and perms set to 700. Now suppose we launch datanode as user "bar". {{DiskChecker#checkDir}} will just look for 700 and not consider the running user, so it will think that the directory is usable. Then, it would experience an I/O error later whenever the process first tries to use that directory. I'll file a separate jira for this. > DiskChecker cannot reliably detect an inaccessible disk on Windows with NTFS > ACLs > --------------------------------------------------------------------------------- > > Key: HADOOP-8973 > URL: https://issues.apache.org/jira/browse/HADOOP-8973 > Project: Hadoop Common > Issue Type: Bug > Components: util > Affects Versions: trunk-win > Reporter: Chris Nauroth > Assignee: Chris Nauroth > Attachments: HADOOP-8973-branch-trunk-win.patch > > > DiskChecker.checkDir uses File.canRead, File.canWrite, and File.canExecute to > check if a directory is inaccessible. These APIs are not reliable on Windows > with NTFS ACLs due to a known JVM bug. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira