Repository: hadoop Updated Branches: refs/heads/branch-2.8 80bb9ffb6 -> ce902fc80
HDFS-10729. Improve log message for edit loading failures caused by FS limit checks. Contributed by Wei-Chiu Chuang. (cherry picked from commit 01721dd88ee532d20eda841254437da4dfd69db5) Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/ce902fc8 Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/ce902fc8 Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/ce902fc8 Branch: refs/heads/branch-2.8 Commit: ce902fc80ae60df4ab50bc421002ccd7078fd571 Parents: 80bb9ff Author: Kihwal Lee <kih...@apache.org> Authored: Wed Aug 31 14:12:02 2016 -0500 Committer: Kihwal Lee <kih...@apache.org> Committed: Wed Aug 31 14:12:02 2016 -0500 ---------------------------------------------------------------------- .../hadoop/hdfs/server/namenode/FSDirWriteFileOp.java | 12 ++++++++---- 1 file changed, 8 insertions(+), 4 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/ce902fc8/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java index fea119a..69a4150 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirWriteFileOp.java @@ -40,6 +40,7 @@ import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.EncryptionZone; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.protocol.FSLimitException; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.QuotaExceededException; @@ -507,10 +508,13 @@ class FSDirWriteFileOp { return newNode; } } catch (IOException e) { - if(NameNode.stateChangeLog.isDebugEnabled()) { - NameNode.stateChangeLog.debug( - "DIR* FSDirectory.unprotectedAddFile: exception when add " - + existing.getPath() + " to the file system", e); + NameNode.stateChangeLog.warn( + "DIR* FSDirectory.unprotectedAddFile: exception when add " + existing + .getPath() + " to the file system", e); + if (e instanceof FSLimitException.MaxDirectoryItemsExceededException) { + NameNode.stateChangeLog.warn("Please increase " + + "dfs.namenode.fs-limits.max-directory-items and make it " + + "consistent across all NameNodes."); } } return null; --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org