[jira] [Created] (HDFS-11955) Ozone: Set proper parameter default values for listBuckets http request
Weiwei Yang created HDFS-11955: -- Summary: Ozone: Set proper parameter default values for listBuckets http request Key: HDFS-11955 URL: https://issues.apache.org/jira/browse/HDFS-11955 Project: Hadoop HDFS Issue Type: Sub-task Components: ozone Reporter: Weiwei Yang Assignee: Weiwei Yang HDFS-11779 implements the listBuckets function in ozone server side, the API supports several parameters, startKey, count and prefix. But both of them are optional for the client side rest API. This jira is to make sure we set proper default values in the http request if they are not explicitly set by users. -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-11954) WritableRpcEngine.Invocation#toString should handle null parameters
John Zhuge created HDFS-11954: - Summary: WritableRpcEngine.Invocation#toString should handle null parameters Key: HDFS-11954 URL: https://issues.apache.org/jira/browse/HDFS-11954 Project: Hadoop HDFS Issue Type: Bug Components: ipc Affects Versions: 2.8.0 Reporter: John Zhuge Assignee: John Zhuge Priority: Minor WritableRpcEngine.Invocation#toString will throw NPE if {{parameters}} is null: {code:title=WritableRpcEngine.Invocation#toString} buffer.append(methodName); buffer.append("("); for (int i = 0; i < parameters.length; i++) { if (i != 0) buffer.append(", "); buffer.append(parameters[i]); } buffer.append(")"); {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/ [Jun 7, 2017 4:21:07 PM] (varunsaxena) YARN-6604. Allow metric TTL for Application table to be specified [Jun 7, 2017 8:34:47 AM] (brahma) MAPREDUCE-6676. NNBench should Throw IOException when rename and delete [Jun 7, 2017 8:41:06 PM] (Carlo Curino) YARN-6634. [API] Refactor ResourceManager WebServices to make API [Jun 7, 2017 8:54:52 PM] (liuml07) HADOOP-14500. Azure: [Jun 7, 2017 10:52:52 PM] (jzhuge) HDFS-11861. ipc.Client.Connection#sendRpcRequest should log request -1 overall The following subsystems voted -1: compile mvninstall unit The following subsystems voted -1 but were configured to be filtered/ignored: cc javac The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: Failed junit tests : hadoop.fs.sftp.TestSFTPFileSystem hadoop.hdfs.server.balancer.TestBalancer hadoop.hdfs.TestDFSStripedOutputStreamWithFailure180 hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer hadoop.hdfs.server.namenode.TestNameNodeMetadataConsistency hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting hadoop.hdfs.TestDFSStripedOutputStreamWithFailure010 hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 hadoop.hdfs.web.TestWebHdfsTimeouts hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy hadoop.yarn.server.nodemanager.recovery.TestNMLeveldbStateStoreService hadoop.yarn.server.nodemanager.TestNodeManagerShutdown hadoop.yarn.server.timeline.TestRollingLevelDB hadoop.yarn.server.timeline.TestTimelineDataManager hadoop.yarn.server.timeline.TestLeveldbTimelineStore hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore hadoop.yarn.server.TestMiniYarnClusterNodeUtilization hadoop.yarn.server.TestContainerManagerSecurity hadoop.yarn.client.api.impl.TestAMRMClient hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore hadoop.yarn.applications.distributedshell.TestDistributedShell hadoop.mapred.TestShuffleHandler hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService Timed out junit tests : org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache org.apache.hadoop.yarn.server.resourcemanager.TestRMStoreCommands org.apache.hadoop.yarn.server.resourcemanager.recovery.TestZKRMStateStore org.apache.hadoop.yarn.server.resourcemanager.TestReservationSystemWithRMHA org.apache.hadoop.yarn.server.resourcemanager.TestSubmitApplicationWithRMHA org.apache.hadoop.yarn.server.resourcemanager.TestKillApplicationWithRMHA mvninstall: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-mvninstall-root.txt [492K] compile: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt [20K] cc: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt [20K] javac: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt [20K] unit: https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-assemblies.txt [4.0K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt [140K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt [288K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-nodemanager.txt [56K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-applicationhistoryservice.txt [52K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt [72K] https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt [324K]
[jira] [Created] (HDFS-11953) Remove Guava v21 usage from HDFS-7240
Xiaoyu Yao created HDFS-11953: - Summary: Remove Guava v21 usage from HDFS-7240 Key: HDFS-11953 URL: https://issues.apache.org/jira/browse/HDFS-11953 Project: Hadoop HDFS Issue Type: Sub-task Affects Versions: HDFS-7240 Reporter: Xiaoyu Yao Assignee: Xiaoyu Yao Due to HADOOP-14386, Guava on trunk has been reverted back from v21 to v11.0.2. There are only two usage currently in HDFS-7240 which prevents us from mergING trunk. This ticket is opened to change them using alternative methods from v11.0.2 so that we can merge trunk. -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/428/ [Jun 7, 2017 4:21:07 PM] (varunsaxena) YARN-6604. Allow metric TTL for Application table to be specified [Jun 7, 2017 8:34:47 AM] (brahma) MAPREDUCE-6676. NNBench should Throw IOException when rename and delete [Jun 7, 2017 8:41:06 PM] (Carlo Curino) YARN-6634. [API] Refactor ResourceManager WebServices to make API [Jun 7, 2017 8:54:52 PM] (liuml07) HADOOP-14500. Azure: [Jun 7, 2017 10:52:52 PM] (jzhuge) HDFS-11861. ipc.Client.Connection#sendRpcRequest should log request -1 overall The following subsystems voted -1: findbugs unit The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: FindBugs : module:hadoop-common-project/hadoop-minikdc Possible null pointer dereference in org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called method Dereferenced at MiniKdc.java:org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called method Dereferenced at MiniKdc.java:[line 368] FindBugs : module:hadoop-common-project/hadoop-auth org.apache.hadoop.security.authentication.server.MultiSchemeAuthenticationHandler.authenticate(HttpServletRequest, HttpServletResponse) makes inefficient use of keySet iterator instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:of keySet iterator instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:[line 192] FindBugs : module:hadoop-common-project/hadoop-common org.apache.hadoop.crypto.CipherSuite.setUnknownValue(int) unconditionally sets the field unknownValue At CipherSuite.java:unknownValue At CipherSuite.java:[line 44] org.apache.hadoop.crypto.CryptoProtocolVersion.setUnknownValue(int) unconditionally sets the field unknownValue At CryptoProtocolVersion.java:unknownValue At CryptoProtocolVersion.java:[line 67] Possible null pointer dereference in org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of called method Dereferenced at FileUtil.java:org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of called method Dereferenced at FileUtil.java:[line 118] Possible null pointer dereference in org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, File, Path, File) due to return value of called method Dereferenced at RawLocalFileSystem.java:org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, File, Path, File) due to return value of called method Dereferenced at RawLocalFileSystem.java:[line 387] Return value of org.apache.hadoop.fs.permission.FsAction.or(FsAction) ignored, but method has no side effect At FTPFileSystem.java:but method has no side effect At FTPFileSystem.java:[line 421] Useless condition:lazyPersist == true at this point At CommandWithDestination.java:[line 502] org.apache.hadoop.io.DoubleWritable.compareTo(DoubleWritable) incorrectly handles double value At DoubleWritable.java: At DoubleWritable.java:[line 78] org.apache.hadoop.io.DoubleWritable$Comparator.compare(byte[], int, int, byte[], int, int) incorrectly handles double value At DoubleWritable.java:int) incorrectly handles double value At DoubleWritable.java:[line 97] org.apache.hadoop.io.FloatWritable.compareTo(FloatWritable) incorrectly handles float value At FloatWritable.java: At FloatWritable.java:[line 71] org.apache.hadoop.io.FloatWritable$Comparator.compare(byte[], int, int, byte[], int, int) incorrectly handles float value At FloatWritable.java:int) incorrectly handles float value At FloatWritable.java:[line 89] Possible null pointer dereference in org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return value of called method Dereferenced at IOUtils.java:org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return value of called method Dereferenced at IOUtils.java:[line 351] org.apache.hadoop.io.erasurecode.ECSchema.toString() makes inefficient use of keySet iterator instead of entrySet iterator At ECSchema.java:keySet iterator instead of entrySet iterator At ECSchema.java:[line 193] Possible bad parsing of shift operation in org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:operation in org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:[line 398] org.apache.hadoop.metrics2.lib.DefaultMetricsFactory.setInstance(MutableMetricsFactory) unconditionally sets the field mmfImpl At DefaultMetricsFactory.java:mmfImpl At DefaultMetricsFactory.java:[line 49] org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.setMiniClusterMode(boolean)
[jira] [Created] (HDFS-11952) Ozone: Fix regression TestContainerSQLCli#testConvertContainerDB
Weiwei Yang created HDFS-11952: -- Summary: Ozone: Fix regression TestContainerSQLCli#testConvertContainerDB Key: HDFS-11952 URL: https://issues.apache.org/jira/browse/HDFS-11952 Project: Hadoop HDFS Issue Type: Sub-task Components: ozone Reporter: Weiwei Yang Assignee: Weiwei Yang TestContainerSQLCli#testConvertContainerDB is failing since HDFS-11568. Error message: {noformat} 2017-06-08 08:21:47,653 [main] ERROR - DB path not exist:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/4/MiniOzoneCluster1113d40f-586f-4914-9ac4-a37c1a3a561d/05bdadbc-1e60-46e0-bf57-efc4f21f2e7e/scm/container.db ... java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at org.apache.hadoop.ozone.scm.TestContainerSQLCli.testConvertContainerDB(TestContainerSQLCli.java:255) {noformat} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-11951) Ozone
Weiwei Yang created HDFS-11951: -- Summary: Ozone Key: HDFS-11951 URL: https://issues.apache.org/jira/browse/HDFS-11951 Project: Hadoop HDFS Issue Type: Sub-task Reporter: Weiwei Yang -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-11950) Disable libhdfs zerocopy test on Mac
John Zhuge created HDFS-11950: - Summary: Disable libhdfs zerocopy test on Mac Key: HDFS-11950 URL: https://issues.apache.org/jira/browse/HDFS-11950 Project: Hadoop HDFS Issue Type: Test Components: libhdfs Affects Versions: 2.8.0 Reporter: John Zhuge Priority: Minor Since libhdfs zerocopy test is expected to fail on Mac, just disable it. {noformat} [exec] Test project /Users/jzhuge/hadoop/hadoop-hdfs-project/hadoop-hdfs-native-client/target [exec] Start 1: test_test_libhdfs_threaded_hdfs_static [exec] 1/3 Test #1: test_test_libhdfs_threaded_hdfs_static ... Passed 9.73 sec [exec] Start 2: test_test_libhdfs_zerocopy_hdfs_static [exec] 2/3 Test #2: test_test_libhdfs_zerocopy_hdfs_static ...***Failed 6.56 sec [exec] Start 3: test_test_native_mini_dfs [exec] Errors while running CTest [exec] 3/3 Test #3: test_test_native_mini_dfs Passed 7.45 sec [exec] [exec] 67% tests passed, 1 tests failed out of 3 [exec] [exec] Total Test time (real) = 23.74 sec [exec] [exec] The following tests FAILED: [exec] 2 - test_test_libhdfs_zerocopy_hdfs_static (Failed) {noformat} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org