[jira] [Commented] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13741995#comment-13741995 ] Hadoop QA commented on HADOOP-9877: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12598358/HADOOP-9877.v2.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.fs.viewfs.TestFcMainOperationsLocalFs org.apache.hadoop.fs.TestFsShellReturnCode org.apache.hadoop.fs.TestGlobPaths {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/2993//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/2993//console This message is automatically generated. hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Binglin Chang updated HADOOP-9877: -- Attachment: HADOOP-9877.v3.patch Add code to explicitly set the path of FileStatus like the original code. hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch, HADOOP-9877.v3.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742082#comment-13742082 ] Hadoop QA commented on HADOOP-9877: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12598391/HADOOP-9877.v3.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs: org.apache.hadoop.fs.TestFsShellReturnCode {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/2994//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/2994//console This message is automatically generated. hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch, HADOOP-9877.v3.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9865) FileContext.globStatus() has a regression with respect to relative path
[ https://issues.apache.org/jira/browse/HADOOP-9865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742092#comment-13742092 ] Hudson commented on HADOOP-9865: SUCCESS: Integrated in Hadoop-Yarn-trunk #303 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/303/]) HADOOP-9865. FileContext#globStatus has a regression with respect to relative path. (Contributed by Chaun Lin) (cmccabe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514531) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestGlobPaths.java FileContext.globStatus() has a regression with respect to relative path --- Key: HADOOP-9865 URL: https://issues.apache.org/jira/browse/HADOOP-9865 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.3.0 Reporter: Chuan Liu Assignee: Chuan Liu Fix For: 2.3.0 Attachments: HADOOP-9865-demo.patch, HADOOP-9865-trunk.2.patch, HADOOP-9865-trunk.3.patch, HADOOP-9865-trunk.patch I discovered the problem when running unit test TestMRJobClient on Windows. The cause is indirect in this case. In the unit test, we try to launch a job and list its status. The job failed, and caused the list command get a result of 0, which triggered the unit test assert. From the log and debug, the job failed because we failed to create the Jar with classpath (see code around {{FileUtil.createJarWithClassPath}}) in {{ContainerLaunch}}. This is a Windows specific step right now; so the test still passes on Linux. This step failed because we passed in a relative path to {{FileContext.globStatus()}} in {{FileUtil.createJarWithClassPath}}. The relevant log looks like the following. {noformat} 2013-08-12 16:12:05,937 WARN [ContainersLauncher #0] launcher.ContainerLaunch (ContainerLaunch.java:call(270)) - Failed to launch container. org.apache.hadoop.HadoopIllegalArgumentException: Path is relative at org.apache.hadoop.fs.Path.checkNotRelative(Path.java:74) at org.apache.hadoop.fs.FileContext.getFSofPath(FileContext.java:304) at org.apache.hadoop.fs.Globber.schemeFromPath(Globber.java:107) at org.apache.hadoop.fs.Globber.glob(Globber.java:128) at org.apache.hadoop.fs.FileContext$Util.globStatus(FileContext.java:1908) at org.apache.hadoop.fs.FileUtil.createJarWithClassPath(FileUtil.java:1247) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.sanitizeEnv(ContainerLaunch.java:679) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:232) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:1) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {noformat} I think this is a regression from HADOOP-9817. I modified some code and the unit test passed. (See the attached patch.) However, I think the impact is larger. I will add some unit tests to verify the behavior, and work on a more complete fix. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9865) FileContext.globStatus() has a regression with respect to relative path
[ https://issues.apache.org/jira/browse/HADOOP-9865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742207#comment-13742207 ] Hudson commented on HADOOP-9865: SUCCESS: Integrated in Hadoop-Hdfs-trunk #1493 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1493/]) HADOOP-9865. FileContext#globStatus has a regression with respect to relative path. (Contributed by Chaun Lin) (cmccabe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514531) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestGlobPaths.java FileContext.globStatus() has a regression with respect to relative path --- Key: HADOOP-9865 URL: https://issues.apache.org/jira/browse/HADOOP-9865 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.3.0 Reporter: Chuan Liu Assignee: Chuan Liu Fix For: 2.3.0 Attachments: HADOOP-9865-demo.patch, HADOOP-9865-trunk.2.patch, HADOOP-9865-trunk.3.patch, HADOOP-9865-trunk.patch I discovered the problem when running unit test TestMRJobClient on Windows. The cause is indirect in this case. In the unit test, we try to launch a job and list its status. The job failed, and caused the list command get a result of 0, which triggered the unit test assert. From the log and debug, the job failed because we failed to create the Jar with classpath (see code around {{FileUtil.createJarWithClassPath}}) in {{ContainerLaunch}}. This is a Windows specific step right now; so the test still passes on Linux. This step failed because we passed in a relative path to {{FileContext.globStatus()}} in {{FileUtil.createJarWithClassPath}}. The relevant log looks like the following. {noformat} 2013-08-12 16:12:05,937 WARN [ContainersLauncher #0] launcher.ContainerLaunch (ContainerLaunch.java:call(270)) - Failed to launch container. org.apache.hadoop.HadoopIllegalArgumentException: Path is relative at org.apache.hadoop.fs.Path.checkNotRelative(Path.java:74) at org.apache.hadoop.fs.FileContext.getFSofPath(FileContext.java:304) at org.apache.hadoop.fs.Globber.schemeFromPath(Globber.java:107) at org.apache.hadoop.fs.Globber.glob(Globber.java:128) at org.apache.hadoop.fs.FileContext$Util.globStatus(FileContext.java:1908) at org.apache.hadoop.fs.FileUtil.createJarWithClassPath(FileUtil.java:1247) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.sanitizeEnv(ContainerLaunch.java:679) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:232) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:1) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {noformat} I think this is a regression from HADOOP-9817. I modified some code and the unit test passed. (See the attached patch.) However, I think the impact is larger. I will add some unit tests to verify the behavior, and work on a more complete fix. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9868) Server must not advertise kerberos realm
[ https://issues.apache.org/jira/browse/HADOOP-9868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742201#comment-13742201 ] Hudson commented on HADOOP-9868: SUCCESS: Integrated in Hadoop-Hdfs-trunk #1493 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1493/]) HADOOP-9868. Server must not advertise kerberos realm. Contributed by Daryn Sharp. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514448) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SaslRpcServer.java Server must not advertise kerberos realm Key: HADOOP-9868 URL: https://issues.apache.org/jira/browse/HADOOP-9868 Project: Hadoop Common Issue Type: Bug Components: ipc Affects Versions: 3.0.0, 2.1.1-beta Reporter: Daryn Sharp Assignee: Daryn Sharp Priority: Blocker Fix For: 3.0.0, 2.1.1-beta Attachments: HADOOP-9868.patch HADOOP-9789 broke kerberos authentication by making the RPC server advertise the kerberos service principal realm. SASL clients and servers do not support specifying a realm, so it must be removed from the advertisement. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9865) FileContext.globStatus() has a regression with respect to relative path
[ https://issues.apache.org/jira/browse/HADOOP-9865?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742279#comment-13742279 ] Hudson commented on HADOOP-9865: FAILURE: Integrated in Hadoop-Mapreduce-trunk #1520 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1520/]) HADOOP-9865. FileContext#globStatus has a regression with respect to relative path. (Contributed by Chaun Lin) (cmccabe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514531) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/Globber.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/TestGlobPaths.java FileContext.globStatus() has a regression with respect to relative path --- Key: HADOOP-9865 URL: https://issues.apache.org/jira/browse/HADOOP-9865 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.3.0 Reporter: Chuan Liu Assignee: Chuan Liu Fix For: 2.3.0 Attachments: HADOOP-9865-demo.patch, HADOOP-9865-trunk.2.patch, HADOOP-9865-trunk.3.patch, HADOOP-9865-trunk.patch I discovered the problem when running unit test TestMRJobClient on Windows. The cause is indirect in this case. In the unit test, we try to launch a job and list its status. The job failed, and caused the list command get a result of 0, which triggered the unit test assert. From the log and debug, the job failed because we failed to create the Jar with classpath (see code around {{FileUtil.createJarWithClassPath}}) in {{ContainerLaunch}}. This is a Windows specific step right now; so the test still passes on Linux. This step failed because we passed in a relative path to {{FileContext.globStatus()}} in {{FileUtil.createJarWithClassPath}}. The relevant log looks like the following. {noformat} 2013-08-12 16:12:05,937 WARN [ContainersLauncher #0] launcher.ContainerLaunch (ContainerLaunch.java:call(270)) - Failed to launch container. org.apache.hadoop.HadoopIllegalArgumentException: Path is relative at org.apache.hadoop.fs.Path.checkNotRelative(Path.java:74) at org.apache.hadoop.fs.FileContext.getFSofPath(FileContext.java:304) at org.apache.hadoop.fs.Globber.schemeFromPath(Globber.java:107) at org.apache.hadoop.fs.Globber.glob(Globber.java:128) at org.apache.hadoop.fs.FileContext$Util.globStatus(FileContext.java:1908) at org.apache.hadoop.fs.FileUtil.createJarWithClassPath(FileUtil.java:1247) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.sanitizeEnv(ContainerLaunch.java:679) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:232) at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:1) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {noformat} I think this is a regression from HADOOP-9817. I modified some code and the unit test passed. (See the attached patch.) However, I think the impact is larger. I will add some unit tests to verify the behavior, and work on a more complete fix. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9868) Server must not advertise kerberos realm
[ https://issues.apache.org/jira/browse/HADOOP-9868?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742273#comment-13742273 ] Hudson commented on HADOOP-9868: FAILURE: Integrated in Hadoop-Mapreduce-trunk #1520 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1520/]) HADOOP-9868. Server must not advertise kerberos realm. Contributed by Daryn Sharp. (kihwal: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514448) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SaslRpcServer.java Server must not advertise kerberos realm Key: HADOOP-9868 URL: https://issues.apache.org/jira/browse/HADOOP-9868 Project: Hadoop Common Issue Type: Bug Components: ipc Affects Versions: 3.0.0, 2.1.1-beta Reporter: Daryn Sharp Assignee: Daryn Sharp Priority: Blocker Fix For: 3.0.0, 2.1.1-beta Attachments: HADOOP-9868.patch HADOOP-9789 broke kerberos authentication by making the RPC server advertise the kerberos service principal realm. SASL clients and servers do not support specifying a realm, so it must be removed from the advertisement. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742357#comment-13742357 ] Sanjay Radia commented on HADOOP-9880: -- Daryn, your hack is slightly more appealing. However, the client-side does know how to deal with StandbyException (ie it tries on the other side). So we need to fix the client side to catch the InvalidToken unwrap the cause and then retry. BTW HDFS-3083 has a test and we need to run that test against this one verify that we have not regressed. SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742387#comment-13742387 ] Elizabeth Thomas commented on HADOOP-9745: -- [~j...@nanthrax.net] I checked out the source code from the trunk (http://svn.apache.org/repos/asf/hadoop/common/trunk and revision 1514608) which is the latest successful build from the link https://builds.apache.org/job/Hadoop-trunk-Commit/4280/. On running a mvn install, the following tests fail, which includes the above three tests. === Results : Failed tests: testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence testCachingRelaysResolveQueries(org.apache.hadoop.net.TestStaticMapping): Expected two entries in the map Mapping: cached switch mapping relaying to static mapping with single switch = false(..) testCachingCachesNegativeEntries(org.apache.hadoop.net.TestStaticMapping): Expected two entries in the map Mapping: cached switch mapping relaying to static mapping with single switch = false(..) testResolve(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] testTableCaching(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] testClearingCachedMappings(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] testNormalizeHostName(org.apache.hadoop.net.TestNetUtils): expected:[67.215.65.145] but was:[UnknownHost123] Tests in error: testGracefulFailover(org.apache.hadoop.ha.TestZKFailoverController): test timed out after 25000 milliseconds testChown(org.apache.hadoop.fs.TestFsShellReturnCode): test timed out after 3 milliseconds testChgrp(org.apache.hadoop.fs.TestFsShellReturnCode): test timed out after 3 milliseconds Tests run: 2128, Failures: 9, Errors: 3, Skipped: 71 = Should I checkout the code from elsewhere so that I could run these unit tests successfully? Or is there a pre-requisite/configuration for these tests (like any Unix service or any other apache services like ZooKeeper to be up) to run successfully on a Unix machine? I believe these unit tests should not have any dependencies. Could you pls provide your pointers? TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Binglin Chang updated HADOOP-9877: -- Attachment: HADOOP-9877.v4.patch hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch, HADOOP-9877.v3.patch, HADOOP-9877.v4.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742414#comment-13742414 ] Binglin Chang commented on HADOOP-9877: --- Upload v4 patch, add mock getFileLinkStatus to TestFsShellReturnCode.RawLocalFileSystemExtn hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch, HADOOP-9877.v3.patch, HADOOP-9877.v4.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9866) convert hadoop-auth testcases requiring kerberos to use minikdc
[ https://issues.apache.org/jira/browse/HADOOP-9866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742466#comment-13742466 ] Wei Yan commented on HADOOP-9866: - update a new pacth according [~tucu00] comments. For TestKerberosName, we don't need the krb5.conf. Only set system properties (realm and host) @Before and clear them in @After. convert hadoop-auth testcases requiring kerberos to use minikdc --- Key: HADOOP-9866 URL: https://issues.apache.org/jira/browse/HADOOP-9866 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 2.3.0 Reporter: Alejandro Abdelnur Assignee: Wei Yan Attachments: HADOOP-9866.patch, HADOOP-9866.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9866) convert hadoop-auth testcases requiring kerberos to use minikdc
[ https://issues.apache.org/jira/browse/HADOOP-9866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742485#comment-13742485 ] Hadoop QA commented on HADOOP-9866: --- {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12598504/HADOOP-9866.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 13 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-auth. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/2996//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/2996//console This message is automatically generated. convert hadoop-auth testcases requiring kerberos to use minikdc --- Key: HADOOP-9866 URL: https://issues.apache.org/jira/browse/HADOOP-9866 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 2.3.0 Reporter: Alejandro Abdelnur Assignee: Wei Yan Attachments: HADOOP-9866.patch, HADOOP-9866.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9873) hadoop-env.sh got called multiple times
[ https://issues.apache.org/jira/browse/HADOOP-9873?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Arpit Agarwal updated HADOOP-9873: -- Affects Version/s: 3.0.0 hadoop-env.sh got called multiple times --- Key: HADOOP-9873 URL: https://issues.apache.org/jira/browse/HADOOP-9873 Project: Hadoop Common Issue Type: Bug Components: scripts Affects Versions: 3.0.0 Reporter: Kai Zheng Assignee: Kai Zheng Priority: Minor Attachments: HADOOP-9873.patch Ref. below, it can be seen hadoop-env.sh got called multiple times when running something like 'hadoop-daemon.sh start namenode'. {noformat} [drankye@zkdev ~]$ cd $HADOOP_PREFIX [drankye@zkdev hadoop-3.0.0-SNAPSHOT]$ grep -r hadoop-env * libexec/hadoop-config.sh:if [ -e ${HADOOP_PREFIX}/conf/hadoop-env.sh ]; then libexec/hadoop-config.sh:if [ -f ${HADOOP_CONF_DIR}/hadoop-env.sh ]; then libexec/hadoop-config.sh: . ${HADOOP_CONF_DIR}/hadoop-env.sh sbin/hadoop-daemon.sh:if [ -f ${HADOOP_CONF_DIR}/hadoop-env.sh ]; then sbin/hadoop-daemon.sh: . ${HADOOP_CONF_DIR}/hadoop-env.sh {noformat} Considering the following lines in hadoop-env.sh {code} # Command specific options appended to HADOOP_OPTS when specified export HADOOP_NAMENODE_OPTS=-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS {code} It may end with some redundant result like below when called multiple times. {noformat} HADOOP_NAMENODE_OPTS='-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender -Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender ' {noformat} It's not a big issue for now however it would be better to be clean and avoid this since it can cause the final JAVA command line is very lengthy and hard to read. A possible fix would be to add a flag variable like HADOOP_ENV_INITED in hadoop-env.sh, and then at the beginning of it check the flag. If the flag evaluates true, then return immediately. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9877) hadoop fsshell can not ls .snapshot dir after HADOOP-9817
[ https://issues.apache.org/jira/browse/HADOOP-9877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742557#comment-13742557 ] Hadoop QA commented on HADOOP-9877: --- {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12598498/HADOOP-9877.v4.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 3 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. The javadoc tool did not generate any warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 1.3.9) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-common-project/hadoop-common hadoop-hdfs-project/hadoop-hdfs. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/2995//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/2995//console This message is automatically generated. hadoop fsshell can not ls .snapshot dir after HADOOP-9817 - Key: HADOOP-9877 URL: https://issues.apache.org/jira/browse/HADOOP-9877 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Binglin Chang Assignee: Binglin Chang Attachments: HADOOP-9877.v1.patch, HADOOP-9877.v2.patch, HADOOP-9877.v3.patch, HADOOP-9877.v4.patch {code} decster:~/hadoop bin/hadoop fs -ls /foo/.snapshot 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/) 13/08/16 01:17:22 INFO hdfs.DFSClient: + listPath(/foo) ls: `/foo/.snapshot': No such file or directory {code} HADOOP-9817 refactor some globStatus code, but forgot to handle special case that .snapshot dir is not show up in listStatus but exists, so we need to explicitly check path existence using getFileStatus, rather than depending on listStatus results. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9742) TestTableMapping test fails
[ https://issues.apache.org/jira/browse/HADOOP-9742?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742565#comment-13742565 ] Elizabeth Thomas commented on HADOOP-9742: -- [~jbonofre] I still these errors while doing a mvn install on the source code checked out from trunk. Could you pls help here on how you resolved them? Any service to be running on Ubuntu to get these test cases to pass? TestTableMapping test fails --- Key: HADOOP-9742 URL: https://issues.apache.org/jira/browse/HADOOP-9742 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testResolve(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] - testTableCaching(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] - testClearingCachedMappings(org.apache.hadoop.net.TestTableMapping): expected:/[rack1] but was:/[default-rack] -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742562#comment-13742562 ] Elizabeth Thomas commented on HADOOP-9745: -- Oops, I included the wrong resolver. Including the correct resolver [~jbonofre]. Jean, could you pls provide pointers? I still see these failure in test cases when I do a mvn install on the source code. TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9843) Backport TestDiskChecker to branch-1.
[ https://issues.apache.org/jira/browse/HADOOP-9843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-9843: -- Target Version/s: 1.3.0 (was: 1-win, 1.3.0) Affects Version/s: (was: 1-win) bq. I found TestDiskChecker have already been backported so the affects/target version may be only branch-1(1.3.0). Yes, you're correct that it's already in branch-1-win. I had been looking at an out-of-date fork. I've changed the version to just 1.3.0. Thanks! bq. I have never seen the NullPointerException on Linux. This is definitely throwing NPE on every run for me. Can you try running the tests yourself again? I know it appears that the patch passed Jenkins, but I'm really confused about what Jenkins did with this patch. Jenkins generally only handles trunk patches, so I'm not sure how or why it replied with +1. Backport TestDiskChecker to branch-1. - Key: HADOOP-9843 URL: https://issues.apache.org/jira/browse/HADOOP-9843 Project: Hadoop Common Issue Type: Improvement Components: test, util Affects Versions: 1.3.0 Reporter: Chris Nauroth Assignee: Kousuke Saruta Priority: Minor Labels: newbie Attachments: HADOOP-9843.patch In trunk, we have the {{TestDiskChecker}} test suite to cover the code in {{DiskChecker}}. It would be good to backport this test suite to branch-1 and branch-1-win to get coverage of the code in those branches too. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9843) Backport TestDiskChecker to branch-1.
[ https://issues.apache.org/jira/browse/HADOOP-9843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-9843: -- Status: Open (was: Patch Available) Backport TestDiskChecker to branch-1. - Key: HADOOP-9843 URL: https://issues.apache.org/jira/browse/HADOOP-9843 Project: Hadoop Common Issue Type: Improvement Components: test, util Affects Versions: 1.3.0 Reporter: Chris Nauroth Assignee: Kousuke Saruta Priority: Minor Labels: newbie Attachments: HADOOP-9843.patch In trunk, we have the {{TestDiskChecker}} test suite to cover the code in {{DiskChecker}}. It would be good to backport this test suite to branch-1 and branch-1-win to get coverage of the code in those branches too. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9743) TestStaticMapping test fails
[ https://issues.apache.org/jira/browse/HADOOP-9743?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742578#comment-13742578 ] Elizabeth Thomas commented on HADOOP-9743: -- [~jbonofre] I still these errors while doing a mvn install on the source code checked out from trunk. Could you pls provide your pointers on how to resolve? Any patch to be applied? TestStaticMapping test fails Key: HADOOP-9743 URL: https://issues.apache.org/jira/browse/HADOOP-9743 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testCachingRelaysResolveQueries(org.apache.hadoop.net.TestStaticMapping): Expected two entries in the map Mapping: cached switch mapping relaying to static mapping with single switch = false(..) - testCachingCachesNegativeEntries(org.apache.hadoop.net.TestStaticMapping): Expected two entries in the map Mapping: cached switch mapping relaying to static mapping with single switch = false(..) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742613#comment-13742613 ] Kihwal Lee commented on HADOOP-9880: bq. However, the client-side does know how to deal with StandbyException (ie it tries on the other side). So we need to fix the client side to catch the InvalidToken unwrap the cause and then retry. Isn't this patch already unwrap InvalidToken from server and throw the cause if it is set? So I thought clients would get StandbyException. Please correct me if I am wrong. I've deployed a secure HA cluster with this patch and it appears to be working. SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742627#comment-13742627 ] Suresh Srinivas commented on HADOOP-9745: - bq. You closed bunch of bugs as fixed. Is it because the issues you observed no longer happen? Earlier my question was not answered. Now that, I understand the reason for closing the issues, the resolution is incorrect. It should be not a problem. I am updating the resolution. If this issue, based on discussions that is going on, is found to be a real issue, please re-open the jira again. TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suresh Srinivas resolved HADOOP-9745. - Resolution: Not A Problem TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Reopened] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suresh Srinivas reopened HADOOP-9745: - TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Comment Edited] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742627#comment-13742627 ] Suresh Srinivas edited comment on HADOOP-9745 at 8/16/13 9:42 PM: -- bq. You closed bunch of bugs as fixed. Is it because the issues you observed no longer happen? Earlier my question was not answered. Now that I understand the reason for closing the issues, the resolution is incorrect. It should be not a problem. I am updating the resolution. If this issue, based on discussions that is going on, is found to be a real issue, please re-open the jira again. was (Author: sureshms): bq. You closed bunch of bugs as fixed. Is it because the issues you observed no longer happen? Earlier my question was not answered. Now that, I understand the reason for closing the issues, the resolution is incorrect. It should be not a problem. I am updating the resolution. If this issue, based on discussions that is going on, is found to be a real issue, please re-open the jira again. TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742636#comment-13742636 ] Elizabeth Thomas commented on HADOOP-9745: -- [~sureshms] and [~jbonofre] I am attempting to build the Hadoop source code and run all the unit tests before I get started to work on the trivial hadoop issues. Could you pls help with pointers on how to get the mvn test or mvn install to run successfully for the source-code checked out from trunk, ignoring these errors/failures? Are these failure scenarios to be ignored? TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9745) TestZKFailoverController test fails
[ https://issues.apache.org/jira/browse/HADOOP-9745?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742639#comment-13742639 ] Elizabeth Thomas commented on HADOOP-9745: -- As given in the http://wiki.apache.org/hadoop/HowToContribute section, I would like to run all the unit test cases before constructing my patch. Pls help with your insights. TestZKFailoverController test fails --- Key: HADOOP-9745 URL: https://issues.apache.org/jira/browse/HADOOP-9745 Project: Hadoop Common Issue Type: Bug Components: test Affects Versions: 3.0.0 Reporter: Jean-Baptiste Onofré - testGracefulFailoverFailBecomingActive(org.apache.hadoop.ha.TestZKFailoverController): Did not fail to graceful failover when target failed to become active! - testGracefulFailoverFailBecomingStandby(org.apache.hadoop.ha.TestZKFailoverController): expected:1 but was:0 - testGracefulFailoverFailBecomingStandbyAndFailFence(org.apache.hadoop.ha.TestZKFailoverController): Failover should have failed when old node wont fence -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Kihwal Lee updated HADOOP-9880: --- Target Version/s: 2.1.1-beta (was: 2.1.0-beta) SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742651#comment-13742651 ] Sanjay Radia commented on HADOOP-9880: -- We also applied and it works. Yes it is unwrapping the exception - i did not read it carefully last night. We applied the test from HDFS-3083 and that test passes. SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HADOOP-9881) Some questions and possible improvement for MiniKdc/TestMiniKdc
Kai Zheng created HADOOP-9881: - Summary: Some questions and possible improvement for MiniKdc/TestMiniKdc Key: HADOOP-9881 URL: https://issues.apache.org/jira/browse/HADOOP-9881 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Kai Zheng In org.apache.hadoop.minikdc.TestMiniKdc: # In testKeytabGen(), it comments ??principals use \ instead of /??, does this mean the principal must use \ instead of / to use MiniKdc for test cases? If so, should *HADOOP_SECURITY_AUTH_TO_LOCAL* consider this? # In testKerberosLogin(), what’s the meant difference between client login and server login? I see isInitiator option is set true or false respectively, but I’m not sure about that. # Both in client login and server login, why loginContext.login() gets called again in the end? Perhaps loginContext.logout(). # It also considers IBM JDK. Ref current UGI implementation, looks like it needs to set KRB5CCNAME system property and useDefaultCcache option specifically. It’s good to test login using keytab as current provided facility and test does. Is it also possible to test login via ticket cache or how to automatically generate ticket cache with specified principal without execution of kinit? This is important to cover user Kerberos login (with kinit) if possible using MiniKdc. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742666#comment-13742666 ] Sanjay Radia commented on HADOOP-9880: -- +1, i will commit in a few minutes; thanks daryn. SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9848) Create a MiniKDC for use with security testing
[ https://issues.apache.org/jira/browse/HADOOP-9848?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742673#comment-13742673 ] Kai Zheng commented on HADOOP-9848: --- HADOOP-9881 was opened to document some questions and possible improvement for MiniKdc/TestMiniKdc. Would you comment on that? Thanks. Create a MiniKDC for use with security testing -- Key: HADOOP-9848 URL: https://issues.apache.org/jira/browse/HADOOP-9848 Project: Hadoop Common Issue Type: New Feature Components: security, test Reporter: Wei Yan Assignee: Wei Yan Fix For: 2.3.0 Attachments: HADOOP-9848.patch, HADOOP-9848.patch, HADOOP-9848.patch, HADOOP-9848.patch, HADOOP-9848.patch Create a MiniKDC using Apache Directory Server. MiniKDC builds an embedded KDC (key distribution center), and allows to create principals and keytabs on the fly. MiniKDC can be integrated for Hadoop security unit testing. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Zhao updated HADOOP-9880: -- Resolution: Fixed Fix Version/s: 2.1.1-beta Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) I've committed this to trunk, branch-2 and branch-2.1-beta. SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Fix For: 2.1.1-beta Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9421) Convert SASL to use ProtoBuf and provide negotiation capabilities
[ https://issues.apache.org/jira/browse/HADOOP-9421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742686#comment-13742686 ] Hudson commented on HADOOP-9421: SUCCESS: Integrated in Hadoop-trunk-Commit #4285 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/4285/]) HADOOP-9880. SASL changes from HADOOP-9421 breaks Secure HA NN. Contributed by Daryn Sharp. (jing9: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514913) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SaslRpcServer.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/security/token/delegation/DelegationTokenSecretManager.java Convert SASL to use ProtoBuf and provide negotiation capabilities - Key: HADOOP-9421 URL: https://issues.apache.org/jira/browse/HADOOP-9421 Project: Hadoop Common Issue Type: Sub-task Affects Versions: 2.0.3-alpha Reporter: Sanjay Radia Assignee: Daryn Sharp Priority: Blocker Fix For: 3.0.0, 2.1.0-beta, 2.3.0 Attachments: HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421.patch, HADOOP-9421-v2-demo.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9880) SASL changes from HADOOP-9421 breaks Secure HA NN
[ https://issues.apache.org/jira/browse/HADOOP-9880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742687#comment-13742687 ] Hudson commented on HADOOP-9880: SUCCESS: Integrated in Hadoop-trunk-Commit #4285 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/4285/]) HADOOP-9880. SASL changes from HADOOP-9421 breaks Secure HA NN. Contributed by Daryn Sharp. (jing9: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1514913) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SaslRpcServer.java * /hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/security/token/delegation/DelegationTokenSecretManager.java SASL changes from HADOOP-9421 breaks Secure HA NN -- Key: HADOOP-9880 URL: https://issues.apache.org/jira/browse/HADOOP-9880 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.1.0-beta Reporter: Kihwal Lee Assignee: Daryn Sharp Priority: Blocker Fix For: 2.1.1-beta Attachments: HADOOP-9880.patch buildSaslNegotiateResponse() will create a SaslRpcServer with TOKEN auth. When create() is called against it, secretManager.checkAvailableForRead() is called, which fails in HA standby. Thus HA standby nodes cannot be transitioned to active. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9639) truly shared cache for jars (jobjar/libjar)
[ https://issues.apache.org/jira/browse/HADOOP-9639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742703#comment-13742703 ] Omkar Vinit Joshi commented on HADOOP-9639: --- something similar.. YARN-1020 truly shared cache for jars (jobjar/libjar) --- Key: HADOOP-9639 URL: https://issues.apache.org/jira/browse/HADOOP-9639 Project: Hadoop Common Issue Type: New Feature Components: filecache Affects Versions: 2.0.4-alpha Reporter: Sangjin Lee Assignee: Sangjin Lee Currently there is the distributed cache that enables you to cache jars and files so that attempts from the same job can reuse them. However, sharing is limited with the distributed cache because it is normally on a per-job basis. On a large cluster, sometimes copying of jobjars and libjars becomes so prevalent that it consumes a large portion of the network bandwidth, not to speak of defeating the purpose of bringing compute to where data is. This is wasteful because in most cases code doesn't change much across many jobs. I'd like to propose and discuss feasibility of introducing a truly shared cache so that multiple jobs from multiple users can share and cache jars. This JIRA is to open the discussion. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9881) Some questions and possible improvement for MiniKdc/TestMiniKdc
[ https://issues.apache.org/jira/browse/HADOOP-9881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742805#comment-13742805 ] Alejandro Abdelnur commented on HADOOP-9881: on #1, no you use as usual the /, you need to use \\ only if you are using the ApacheDS KeytabDecoder to read a keytab, which you normally will not do. on #2, the test is testing both a client and a server session, this is standard Java GSS on #3, you are correct, it should be login/logout login/logout, thx for catching that on #4, it has not been tested with IBM JDK, code from hadoop-auth was just copied here. it may need to be corrected. Regarding using kinit from the command line, that cannot be done with minikdc because the minikdc starts and stops with the test, so you cannot do a kinit against a kdc does not yet exist. Though, from Hadoop perspective, if you do a LoginContext like the testcases in minikdc, you are not using UGI keytab login, but an existing session, that it is equivalent to a command line kinit, you just have to make sure you you do a Subject.doAs(subject, PrivilegedAction) to test the hadoop code. I think in this JIRA we can take care of #3 and #4. Some questions and possible improvement for MiniKdc/TestMiniKdc --- Key: HADOOP-9881 URL: https://issues.apache.org/jira/browse/HADOOP-9881 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Kai Zheng In org.apache.hadoop.minikdc.TestMiniKdc: # In testKeytabGen(), it comments ??principals use \ instead of /??, does this mean the principal must use \ instead of / to use MiniKdc for test cases? If so, should *HADOOP_SECURITY_AUTH_TO_LOCAL* consider this? # In testKerberosLogin(), what’s the meant difference between client login and server login? I see isInitiator option is set true or false respectively, but I’m not sure about that. # Both in client login and server login, why loginContext.login() gets called again in the end? Perhaps loginContext.logout(). # It also considers IBM JDK. Ref current UGI implementation, looks like it needs to set KRB5CCNAME system property and useDefaultCcache option specifically. It’s good to test login using keytab as current provided facility and test does. Is it also possible to test login via ticket cache or how to automatically generate ticket cache with specified principal without execution of kinit? This is important to cover user Kerberos login (with kinit) if possible using MiniKdc. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HADOOP-9881) Some questions and possible improvement for MiniKdc/TestMiniKdc
[ https://issues.apache.org/jira/browse/HADOOP-9881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13742824#comment-13742824 ] Kai Zheng commented on HADOOP-9881: --- Thanks for your answers. Might I repeat the question. As it can generate keytab and then do keytab login, is it possible to do the similar work as kinit does to automatically generate ticket cache with specified principal? It's useful to test the USER_KERBEROS_OPTIONS for UGI. Some questions and possible improvement for MiniKdc/TestMiniKdc --- Key: HADOOP-9881 URL: https://issues.apache.org/jira/browse/HADOOP-9881 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Kai Zheng In org.apache.hadoop.minikdc.TestMiniKdc: # In testKeytabGen(), it comments ??principals use \ instead of /??, does this mean the principal must use \ instead of / to use MiniKdc for test cases? If so, should *HADOOP_SECURITY_AUTH_TO_LOCAL* consider this? # In testKerberosLogin(), what’s the meant difference between client login and server login? I see isInitiator option is set true or false respectively, but I’m not sure about that. # Both in client login and server login, why loginContext.login() gets called again in the end? Perhaps loginContext.logout(). # It also considers IBM JDK. Ref current UGI implementation, looks like it needs to set KRB5CCNAME system property and useDefaultCcache option specifically. It’s good to test login using keytab as current provided facility and test does. Is it also possible to test login via ticket cache or how to automatically generate ticket cache with specified principal without execution of kinit? This is important to cover user Kerberos login (with kinit) if possible using MiniKdc. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira