[ https://issues.apache.org/jira/browse/HADOOP-1875?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12531869 ]
Hairong Kuang commented on HADOOP-1875: --------------------------------------- By simply reading the code, it seems that LocalDirAllocator does fail over to a different directory when a tmp directory is not writable when allocating a new directory. But if the allocated directory becomes not writable while data is writing, dfs does not handle the error. Christian, could you please publish the failure stack trace? So I am able to pinpoint the error. Thanks. > multiple dfs.client.buffer.dir directories are not treated as alternatives > -------------------------------------------------------------------------- > > Key: HADOOP-1875 > URL: https://issues.apache.org/jira/browse/HADOOP-1875 > Project: Hadoop > Issue Type: Bug > Components: fs > Reporter: Christian Kunz > Assignee: Hairong Kuang > Priority: Blocker > Fix For: 0.15.0 > > > When specifying multiple directories in the value for dfs.client.buffer.dir, > jobs fail when the selected directory does not exist or is not writable. > Correct behaviour should be to create the directory when it does not exist > and fail over to an alternative directory when it is not writable. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.