[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13682893#comment-13682893 ] Suresh Srinivas commented on HDFS-3934: --- Colin, please mark the Fixed Version(s) field as 3.0.0 on committing a patch to trunk. Please do merge this to branch-2.1-beta. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13680667#comment-13680667 ] Colin Patrick McCabe commented on HDFS-3934: Just a quick reminder, if there are no comments on this we'll commit to branch-2 / branch-2.1 in a day or two. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13677293#comment-13677293 ] Colin Patrick McCabe commented on HDFS-3934: [~daryn] I'm going to wait a few days before merging to branch-2, in case you want to take a second look at it. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13674254#comment-13674254 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-Yarn-trunk #230 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/230/]) Remove extra code that code erroneously committed in HDFS-3934 (cmccabe) (Revision 1489083) Add needed file for HDFS-3934 (cmccabe) (Revision 1489068) HDFS-3934. duplicative dfs_hosts entries handled wrong. (cmccabe) (Revision 1489065) Result = FAILURE cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489083 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489068 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489065 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDatanodeRegistration.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13674345#comment-13674345 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-Hdfs-trunk #1420 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1420/]) Remove extra code that code erroneously committed in HDFS-3934 (cmccabe) (Revision 1489083) Add needed file for HDFS-3934 (cmccabe) (Revision 1489068) HDFS-3934. duplicative dfs_hosts entries handled wrong. (cmccabe) (Revision 1489065) Result = FAILURE cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489083 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489068 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489065 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDatanodeRegistration.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13674413#comment-13674413 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-Mapreduce-trunk #1446 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1446/]) Remove extra code that code erroneously committed in HDFS-3934 (cmccabe) (Revision 1489083) Add needed file for HDFS-3934 (cmccabe) (Revision 1489068) HDFS-3934. duplicative dfs_hosts entries handled wrong. (cmccabe) (Revision 1489065) Result = SUCCESS cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489083 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489068 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489065 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDatanodeRegistration.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13673334#comment-13673334 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-trunk-Commit #3840 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/3840/]) HDFS-3934. duplicative dfs_hosts entries handled wrong. (cmccabe) (Revision 1489065) Result = FAILURE cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489065 Files : * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/HostsFileReader.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDatanodeRegistration.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13673340#comment-13673340 ] Colin Patrick McCabe commented on HDFS-3934: I talked to Daryn offline about this and he said he was ok with this going in, though he didn't have time this week to re-review. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13673349#comment-13673349 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-trunk-Commit #3841 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/3841/]) Add needed file for HDFS-3934 (cmccabe) (Revision 1489068) Result = SUCCESS cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489068 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13673375#comment-13673375 ] Hudson commented on HDFS-3934: -- Integrated in Hadoop-trunk-Commit #3843 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/3843/]) Remove extra code that code erroneously committed in HDFS-3934 (cmccabe) (Revision 1489083) Result = SUCCESS cmccabe : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1489083 Files : * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/DatanodeManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/HostFileManager.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDecommission.java duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13670564#comment-13670564 ] Daryn Sharp commented on HDFS-3934: --- Thanks, I'll look again this afternoon. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13670803#comment-13670803 ] Colin Patrick McCabe commented on HDFS-3934: thanks, Daryn. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13669871#comment-13669871 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12585302/HDFS-3934.017.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4455//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4455//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13669880#comment-13669880 ] Aaron T. Myers commented on HDFS-3934: -- Makes sense re: points 8 and 11. I agree with both of the other fixes you made in response to the failing tests, too. +1, the latest patch looks good to me. I suggest you wait a day to see if Daryn has any comments before committing this, though. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch, HDFS-3934.017.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1381#comment-1381 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12584734/HDFS-3934.016.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4440//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4440//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch, HDFS-3934.016.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13665668#comment-13665668 ] Aaron T. Myers commented on HDFS-3934: -- The latest patch looks pretty good to me, Colin. A few comments: # Recommend you add @VisibleForTesting annotation to DataNode#getXferPort. # Recommend you change this exception message to include the full entry string, not just the port part that couldn't be parsed, and mention explicitly that it was the port component that couldn't be parsed: {code} + throw new IOException(invalid number format when parsing + + portStr, e); {code} The fact that it was an invalid number and what the number was will already be contained in the message for the NumberFormatException: # Recommend you add something to this warning message to make it clear that this is expected if using the DN registration name feature, and to make it clear that this was encountered when reading an include or exclude file: {code} +LOG.warn(unknown host + prefix, e); {code} # In HostFileManager.EntrySet#find(Entry), since right after this we uncondtionally return null, you can condense this code into a single {{if (...)}} condition which first checks that {{ceil != null}}: {code} +if (ceil == null) { + return null; +} +if (ceil.getValue().getIdentifier().equals( + toFind.getIdentifier())) { + return ceil.getValue(); +} {code} # In HostFileManager.MutableEntrySet#add(DatanodeID), are we guaranteed that datanodeID.getXferPort() = 0? Perhaps we should assert that? # Perhaps make HostFileManager.EntrySet.index protected? # I see the purpose of delaying the throwing of the errors in HostFileManager#refresh, but you might want to add a comment explaining it, since it's not super obvious. I'd also recommend adding something to the log messages in that method along the lines of failed to read exclude file, continuing to use previous list of excluded nodes to make it clear what happens in this case. # Perhaps I'm missing something, but why have separate classes for EntrySet and MutableEntrySet? The only time we use just the normal EntrySet is for the initial empty sets, so seems like we should just have a single class. # Seems like you should do an s/DataNode/NameNode/g in this comment: {code} +// These entries will be de-duped by the DataNode, since they refer +// to the same IP address + port combo. {code} # I don't think this code is doing anything in TestDecommission#testDuplicateHostEntries: {code} +info = client.datanodeReport(DatanodeReportType.DEAD); {code} # Seems like you could replace all of this code with just three asserts: two calls to Map#contains(...), and one check that Map#size() == 2: {code} +IteratorMap.EntryString, DatanodeInfo iter = +deadByXferAddr.entrySet().iterator(); +boolean foundPort1 = false, foundPort2 = false; +while (iter.hasNext()) { + Map.EntryString, DatanodeInfo entry = iter.next(); + DatanodeInfo dn = entry.getValue(); + if (dn.getXferPort() == port1) { +foundPort1 = true; +iter.remove(); + } else if (dn.getXferPort() == port2) { +foundPort2 = true; +iter.remove(); + } +} +Assert.assertTrue(did not find a dead entry with port + port1, +foundPort1); +Assert.assertTrue(did not find a dead entry with port + port2, +foundPort2); +Assert.assertTrue(deadByXferAddr.isEmpty()); {code} # I like that you make a copy of the Configuration object in testIncludeByRegistrationName, and recommend you do the same in testDuplicateHostsEntries, just to minimize the likelihood of inter-test interference. I'll be +1 once these are addressed. Daryn (or anyone who's intending to review this) - please do so shortly. I'll be committing this soon after Colin posts a patch addressing these comments unless I hear from someone else. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13665972#comment-13665972 ] Colin Patrick McCabe commented on HDFS-3934: points 1, 2, 3, 4, 5, 6, 7, 9, 10, 12: thanks, will fix. Point 8: {{MutableEntrySet}} is used internally in {{HostFileManager}}, but I don't want it to escape to {{DatanodeManager}}. I don't want {{DatanodeManager}] to be able to mutate this set. I was considering using an {{ImmutableMap}}, but unfortunately, {{ImmutableMap#Builder}} chokes when you try to feed it two identical keys with different values. There might be a more elegant way of doing this, but for now, this seemed like a nice way to give DatanodeManager a read-only view without doing a lot of copying. Point 11: I see what you're getting at here, but it's a little trickier than it might seem. If I used {{Map#containsValue}}, I'd have to create a DatanodeInfo object which compared equal to what I was looking for. Then I have to start thinking about what fields {{DatanodeInfo#equals}} looks at, which I kind of wanted to avoid. It would work, but it feels kind of messy (I'd be creating a temporary DatanodeID where a lot of the fields were set to dummy values because {{DatanodeInfo#equals}} doesn't care about them, and I don't either.) So I dunno. Maybe there is a better way to do, but it's not obvious to me. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13666017#comment-13666017 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12584639/HDFS-3934.015.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.TestDistributedFileSystem org.apache.hadoop.hdfs.TestDatanodeRegistration {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4434//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4434//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch, HDFS-3934.015.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13646613#comment-13646613 ] Daryn Sharp commented on HDFS-3934: --- Patch is bigger than I expected, so I'll try to review soon because critical webhdfs issues are dominating my time. Poke me next week in case I forget. I did notice log messages are prefaced with WATERMELON. :) duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13646828#comment-13646828 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12581373/HDFS-3934.014.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.blockmanagement.TestBlocksWithNotEnoughRacks {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4352//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4352//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13646932#comment-13646932 ] Colin Patrick McCabe commented on HDFS-3934: Looks like the test failure is https://issues.apache.org/jira/browse/HDFS-3538. Known flaky test. By the way, this patch is the same as the previous except for the comment change. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch, HDFS-3934.014.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13646023#comment-13646023 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12581212/HDFS-3934.013.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.balancer.TestBalancerWithNodeGroup {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4345//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4345//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13646071#comment-13646071 ] Colin Patrick McCabe commented on HDFS-3934: Test failure looks like https://issues.apache.org/jira/browse/HDFS-4261, unrelated to patch. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch, HDFS-3934.013.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13644721#comment-13644721 ] Colin Patrick McCabe commented on HDFS-3934: * properly handle DatanodeID objects where {{getIpAddr()}} and/or {{getHostName()}} return {{null}}. This fixes the two failing tests. * don't log an error when an include/exclude file is set to the empty string. (this is perfectly acceptable; it just means that we don't have such a file.) * log the name of the include/exclude file we failed to read in our error message. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13644835#comment-13644835 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12581007/HDFS-3934.012.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:red}-1 javadoc{color}. The javadoc tool appears to have generated 2 warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:red}-1 findbugs{color}. The patch appears to introduce 2 new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4336//testReport/ Findbugs warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/4336//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4336//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch, HDFS-3934.012.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13644255#comment-13644255 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12580920/HDFS-3934.011.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:red}-1 javadoc{color}. The javadoc tool appears to have generated 2 warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:red}-1 findbugs{color}. The patch appears to introduce 2 new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.namenode.TestNNThroughputBenchmark org.apache.hadoop.hdfs.TestDatanodeRegistration {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4335//testReport/ Findbugs warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/4335//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4335//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch, HDFS-3934.011.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13643416#comment-13643416 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12580790/HDFS-3934.010.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:red}-1 javac{color:red}. The patch appears to cause the build to fail. Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4330//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch, HDFS-3934.010.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13641848#comment-13641848 ] Daryn Sharp commented on HDFS-3934: --- Quick review just by eyeballing the patch: It'd be nice to continue to use the {{HostsFileReader}} and post-process the result. Otherwise it's a consistency/maintenance to copy-n-paste any new parsing functionality. Why does the reader need to instantiate dummy {{DatanodeID}}? It appears to be for repeatedly making the somewhat fragile assumption that xferAddr is ipAddr+port? If that relationship changes, we've got a problem... Patch appears to have dropped supported for the node's registration name. [~eli] wanted me to maintain that feature in HDFS-3990. If we need to keep it, doing a lookup and a canonical lookup (can trigger another dns lookup) isn't compatible with supporting the reg name. Doing a lookup followed by {{getCanonicalName}} is a bad idea. It does 2 more lookups: hostname - PTR - A so it can resolve CNAMES to IP to hostname. With this change I think it will cause 3 lookups per host. Question about // If no transfer port was specified, we take a guess. Why needed, and what are the ramifications for getting this wrong? Just a display issue? It _feels_ like de-dupping the display should be a bit easier to do w/o changing core node registration logic? duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13632321#comment-13632321 ] Colin Patrick McCabe commented on HDFS-3934: Thinking about this a little more, I think we need to consider an entry in the excludes or includes file which does not contain a port to cover *any* DataNode on that host. If we don't, there is a risk of breaking existing clients with this change (and that is the breakage observed in {{TestStartup}}). duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13632464#comment-13632464 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12578837/HDFS-3934.008.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4257//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4257//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Colin Patrick McCabe Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch, HDFS-3934.008.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628740#comment-13628740 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12578144/HDFS-3934.006.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.TestDecommission org.apache.hadoop.hdfs.server.namenode.TestStartup {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4219//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/4219//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4219//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13629260#comment-13629260 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12578245/HDFS-3934.007.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.namenode.TestStartup {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4222//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4222//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch, HDFS-3934.007.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628171#comment-13628171 ] Todd Lipcon commented on HDFS-3934: --- {code} +this.defaultInfoPort = NetUtils.createSocketAddr( +conf.get(DFSConfigKeys.DFS_DATANODE_HTTPS_ADDRESS_KEY, {code} This doesn't seem right -- the HTTPS port isn't the default unless it's a secure cluster. I'm a little nervous about the additional DNS checks inside getDatanodeListForReport. I think this might end up reverting part of the work done in HDFS-3990. Daryn, can you take a look at this patch? duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628230#comment-13628230 ] Colin Patrick McCabe commented on HDFS-3934: It's true that this does add additional DNS checks for {{dfs.hosts.exclude}} and {{dfs.hosts}}. It should be possible to drop the {{FSNamesystem}} lock while these DNS resolutions are going on. Alternately, we could move the DNS resolutions into {{DatanodeManager#refreshHostsReader}}. This would mean that administrators would have to re-run {{dfsadmin -refreshNodes}} if the DNS mapping changed for a hostname in the include or excludes file. Finally, we could simply tell people who are concerned about DNS resolution time to use IP addresses in the include and exclude files. We'd also have to change the code to avoid doing the reverse DNS lookup in this scenario. I think I like solution #2 the best; what do you think? duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628235#comment-13628235 ] Todd Lipcon commented on HDFS-3934: --- Which one's solution #2? This one? {quote} Alternately, we could move the DNS resolutions into DatanodeManager#refreshHostsReader. This would mean that administrators would have to re-run dfsadmin -refreshNodes if the DNS mapping changed for a hostname in the include or excludes file. {quote} I think that's a reasonable thing... basically, when we read the list, we add an exclude entry to our internal data structure both for the IP address and its canonical hostname, so if a DN registers with either one, we'll reject it. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628553#comment-13628553 ] Todd Lipcon commented on HDFS-3934: --- Quick notes on the patch: - Why are you using the visitor pattern instead of just exposing getters from the reader class? Doesn't seem like there's much you're gaining out of it, since it's just a simple iteration over a list. If you're worried about exposing internal state, you could make the fields of the reader class ImmutableMaps -- it looks like they're atomically replaced by refresh() anyway. - Seems like the refresh code could avoid synchronization until the point where the new maps are swapped in -- otherwise if the disk is slow during refreshNodes, or the DNS is slow, any other operations will end up blocking - New file is missing interface audience and license I'll let Daryn or someone else who knows this area of the code a bit better comment on whether the fix is actually semantically correct. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13628626#comment-13628626 ] Hadoop QA commented on HDFS-3934: - {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12578129/HDFS-3934.005.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:red}-1 release audit{color}. The applied patch generated 1 release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.hdfs.server.namenode.TestStartup {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4217//testReport/ Release audit warnings: https://builds.apache.org/job/PreCommit-HDFS-Build/4217//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4217//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch, HDFS-3934.005.patch, HDFS-3934.006.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13627193#comment-13627193 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12577885/HDFS-3934.003.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4211//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4211//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13627248#comment-13627248 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12577899/HDFS-3934.004.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4212//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4212//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch, HDFS-3934.003.patch, HDFS-3934.004.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HDFS-3934) duplicative dfs_hosts entries handled wrong
[ https://issues.apache.org/jira/browse/HDFS-3934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13626234#comment-13626234 ] Hadoop QA commented on HDFS-3934: - {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12577696/HDFS-3934.002.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HDFS-Build/4207//testReport/ Console output: https://builds.apache.org/job/PreCommit-HDFS-Build/4207//console This message is automatically generated. duplicative dfs_hosts entries handled wrong --- Key: HDFS-3934 URL: https://issues.apache.org/jira/browse/HDFS-3934 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 2.0.1-alpha Reporter: Andy Isaacson Assignee: Andy Isaacson Priority: Minor Attachments: HDFS-3934.001.patch, HDFS-3934.002.patch A dead DN listed in dfs_hosts_allow.txt by IP and in dfs_hosts_exclude.txt by hostname ends up being displayed twice in {{dfsnodelist.jsp?whatNodes=DEAD}} after the NN restarts because {{getDatanodeListForReport}} does not handle such a pseudo-duplicate correctly: # the Remove any nodes we know about from the map loop no longer has the knowledge to remove the spurious entries # the The remaining nodes are ones that are referenced by the hosts files loop does not do hostname lookups, so does not know that the IP and hostname refer to the same host. Relatedly, such an IP-based dfs_hosts entry results in a cosmetic problem in the JSP output: The *Node* column shows :50010 as the nodename, with HTML markup {{a href=http://:50075/browseDirectory.jsp?namenodeInfoPort=50070amp;dir=%2Famp;nnaddr=172.29.97.196:8020; title=172.29.97.216:50010:50010/a}}. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira