[jira] [Created] (HDFS-11955) Ozone: Set proper parameter default values for listBuckets http request

2017-06-08 Thread Weiwei Yang (JIRA)
Weiwei Yang created HDFS-11955:
--

 Summary: Ozone: Set proper parameter default values for 
listBuckets http request
 Key: HDFS-11955
 URL: https://issues.apache.org/jira/browse/HDFS-11955
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ozone
Reporter: Weiwei Yang
Assignee: Weiwei Yang


HDFS-11779 implements the listBuckets function in ozone server side, the API 
supports several parameters, startKey, count and prefix. But both of them are 
optional for the client side rest API. This jira is to make sure we set proper 
default values in the http request if they are not explicitly set by users.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11954) WritableRpcEngine.Invocation#toString should handle null parameters

2017-06-08 Thread John Zhuge (JIRA)
John Zhuge created HDFS-11954:
-

 Summary: WritableRpcEngine.Invocation#toString should handle null 
parameters
 Key: HDFS-11954
 URL: https://issues.apache.org/jira/browse/HDFS-11954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: ipc
Affects Versions: 2.8.0
Reporter: John Zhuge
Assignee: John Zhuge
Priority: Minor


WritableRpcEngine.Invocation#toString will throw NPE if {{parameters}} is null:
{code:title=WritableRpcEngine.Invocation#toString}
  buffer.append(methodName);
  buffer.append("(");
  for (int i = 0; i < parameters.length; i++) {
if (i != 0)
  buffer.append(", ");
buffer.append(parameters[i]);
  }
  buffer.append(")");
{code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le

2017-06-08 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/

[Jun 7, 2017 4:21:07 PM] (varunsaxena) YARN-6604. Allow metric TTL for 
Application table to be specified
[Jun 7, 2017 8:34:47 AM] (brahma) MAPREDUCE-6676. NNBench should Throw 
IOException when rename and delete
[Jun 7, 2017 8:41:06 PM] (Carlo Curino) YARN-6634. [API] Refactor 
ResourceManager WebServices to make API
[Jun 7, 2017 8:54:52 PM] (liuml07) HADOOP-14500. Azure:
[Jun 7, 2017 10:52:52 PM] (jzhuge) HDFS-11861. 
ipc.Client.Connection#sendRpcRequest should log request




-1 overall


The following subsystems voted -1:
compile mvninstall unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc javac


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.fs.sftp.TestSFTPFileSystem 
   hadoop.hdfs.server.balancer.TestBalancer 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure180 
   hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer 
   hadoop.hdfs.server.namenode.TestNameNodeMetadataConsistency 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure010 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy 
   hadoop.yarn.server.nodemanager.recovery.TestNMLeveldbStateStoreService 
   hadoop.yarn.server.nodemanager.TestNodeManagerShutdown 
   hadoop.yarn.server.timeline.TestRollingLevelDB 
   hadoop.yarn.server.timeline.TestTimelineDataManager 
   hadoop.yarn.server.timeline.TestLeveldbTimelineStore 
   hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore 
   hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore 
   
hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer 
   hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer 
   hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore 
   hadoop.yarn.server.TestMiniYarnClusterNodeUtilization 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.client.api.impl.TestAMRMClient 
   hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore 
   hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient 
   hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.mapred.TestShuffleHandler 
   hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService 

Timed out junit tests :

   org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache 
   org.apache.hadoop.yarn.server.resourcemanager.TestRMStoreCommands 
   
org.apache.hadoop.yarn.server.resourcemanager.recovery.TestZKRMStateStore 
   
org.apache.hadoop.yarn.server.resourcemanager.TestReservationSystemWithRMHA 
   
org.apache.hadoop.yarn.server.resourcemanager.TestSubmitApplicationWithRMHA 
   
org.apache.hadoop.yarn.server.resourcemanager.TestKillApplicationWithRMHA 
  

   mvninstall:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-mvninstall-root.txt
  [492K]

   compile:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt
  [20K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt
  [20K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-compile-root.txt
  [20K]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-assemblies.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt
  [140K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [288K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-nodemanager.txt
  [56K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-applicationhistoryservice.txt
  [52K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
  [72K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/339/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-tests.txt
  [324K]
 

[jira] [Created] (HDFS-11953) Remove Guava v21 usage from HDFS-7240

2017-06-08 Thread Xiaoyu Yao (JIRA)
Xiaoyu Yao created HDFS-11953:
-

 Summary: Remove Guava v21 usage from HDFS-7240
 Key: HDFS-11953
 URL: https://issues.apache.org/jira/browse/HDFS-11953
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7240
Reporter: Xiaoyu Yao
Assignee: Xiaoyu Yao


Due to HADOOP-14386, Guava on trunk has been reverted back from v21 to v11.0.2. 
There are only two usage currently in HDFS-7240 which prevents us from mergING 
trunk. This ticket is opened to change them using alternative methods from 
v11.0.2 so that we can merge trunk.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2017-06-08 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/428/

[Jun 7, 2017 4:21:07 PM] (varunsaxena) YARN-6604. Allow metric TTL for 
Application table to be specified
[Jun 7, 2017 8:34:47 AM] (brahma) MAPREDUCE-6676. NNBench should Throw 
IOException when rename and delete
[Jun 7, 2017 8:41:06 PM] (Carlo Curino) YARN-6634. [API] Refactor 
ResourceManager WebServices to make API
[Jun 7, 2017 8:54:52 PM] (liuml07) HADOOP-14500. Azure:
[Jun 7, 2017 10:52:52 PM] (jzhuge) HDFS-11861. 
ipc.Client.Connection#sendRpcRequest should log request




-1 overall


The following subsystems voted -1:
findbugs unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

FindBugs :

   module:hadoop-common-project/hadoop-minikdc 
   Possible null pointer dereference in 
org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called 
method Dereferenced at 
MiniKdc.java:org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value 
of called method Dereferenced at MiniKdc.java:[line 368] 

FindBugs :

   module:hadoop-common-project/hadoop-auth 
   
org.apache.hadoop.security.authentication.server.MultiSchemeAuthenticationHandler.authenticate(HttpServletRequest,
 HttpServletResponse) makes inefficient use of keySet iterator instead of 
entrySet iterator At MultiSchemeAuthenticationHandler.java:of keySet iterator 
instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:[line 
192] 

FindBugs :

   module:hadoop-common-project/hadoop-common 
   org.apache.hadoop.crypto.CipherSuite.setUnknownValue(int) 
unconditionally sets the field unknownValue At CipherSuite.java:unknownValue At 
CipherSuite.java:[line 44] 
   org.apache.hadoop.crypto.CryptoProtocolVersion.setUnknownValue(int) 
unconditionally sets the field unknownValue At 
CryptoProtocolVersion.java:unknownValue At CryptoProtocolVersion.java:[line 67] 
   Possible null pointer dereference in 
org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of 
called method Dereferenced at 
FileUtil.java:org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to 
return value of called method Dereferenced at FileUtil.java:[line 118] 
   Possible null pointer dereference in 
org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, 
File, Path, File) due to return value of called method Dereferenced at 
RawLocalFileSystem.java:org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path,
 File, Path, File) due to return value of called method Dereferenced at 
RawLocalFileSystem.java:[line 387] 
   Return value of org.apache.hadoop.fs.permission.FsAction.or(FsAction) 
ignored, but method has no side effect At FTPFileSystem.java:but method has no 
side effect At FTPFileSystem.java:[line 421] 
   Useless condition:lazyPersist == true at this point At 
CommandWithDestination.java:[line 502] 
   org.apache.hadoop.io.DoubleWritable.compareTo(DoubleWritable) 
incorrectly handles double value At DoubleWritable.java: At 
DoubleWritable.java:[line 78] 
   org.apache.hadoop.io.DoubleWritable$Comparator.compare(byte[], int, int, 
byte[], int, int) incorrectly handles double value At DoubleWritable.java:int) 
incorrectly handles double value At DoubleWritable.java:[line 97] 
   org.apache.hadoop.io.FloatWritable.compareTo(FloatWritable) incorrectly 
handles float value At FloatWritable.java: At FloatWritable.java:[line 71] 
   org.apache.hadoop.io.FloatWritable$Comparator.compare(byte[], int, int, 
byte[], int, int) incorrectly handles float value At FloatWritable.java:int) 
incorrectly handles float value At FloatWritable.java:[line 89] 
   Possible null pointer dereference in 
org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return 
value of called method Dereferenced at 
IOUtils.java:org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) 
due to return value of called method Dereferenced at IOUtils.java:[line 351] 
   org.apache.hadoop.io.erasurecode.ECSchema.toString() makes inefficient 
use of keySet iterator instead of entrySet iterator At ECSchema.java:keySet 
iterator instead of entrySet iterator At ECSchema.java:[line 193] 
   Possible bad parsing of shift operation in 
org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At 
Utils.java:operation in 
org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:[line 
398] 
   
org.apache.hadoop.metrics2.lib.DefaultMetricsFactory.setInstance(MutableMetricsFactory)
 unconditionally sets the field mmfImpl At DefaultMetricsFactory.java:mmfImpl 
At DefaultMetricsFactory.java:[line 49] 
   
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.setMiniClusterMode(boolean) 

[jira] [Created] (HDFS-11952) Ozone: Fix regression TestContainerSQLCli#testConvertContainerDB

2017-06-08 Thread Weiwei Yang (JIRA)
Weiwei Yang created HDFS-11952:
--

 Summary: Ozone: Fix regression 
TestContainerSQLCli#testConvertContainerDB
 Key: HDFS-11952
 URL: https://issues.apache.org/jira/browse/HDFS-11952
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ozone
Reporter: Weiwei Yang
Assignee: Weiwei Yang


TestContainerSQLCli#testConvertContainerDB is failing since HDFS-11568. Error 
message:
{noformat}
2017-06-08 08:21:47,653 [main] ERROR  - DB path not 
exist:/testptch/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/test/data/4/MiniOzoneCluster1113d40f-586f-4914-9ac4-a37c1a3a561d/05bdadbc-1e60-46e0-bf57-efc4f21f2e7e/scm/container.db
...
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.ozone.scm.TestContainerSQLCli.testConvertContainerDB(TestContainerSQLCli.java:255)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11951) Ozone

2017-06-08 Thread Weiwei Yang (JIRA)
Weiwei Yang created HDFS-11951:
--

 Summary: Ozone
 Key: HDFS-11951
 URL: https://issues.apache.org/jira/browse/HDFS-11951
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Weiwei Yang






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11950) Disable libhdfs zerocopy test on Mac

2017-06-08 Thread John Zhuge (JIRA)
John Zhuge created HDFS-11950:
-

 Summary: Disable libhdfs zerocopy test on Mac
 Key: HDFS-11950
 URL: https://issues.apache.org/jira/browse/HDFS-11950
 Project: Hadoop HDFS
  Issue Type: Test
  Components: libhdfs
Affects Versions: 2.8.0
Reporter: John Zhuge
Priority: Minor


Since libhdfs zerocopy test is expected to fail on Mac, just disable it.
{noformat}
 [exec] Test project 
/Users/jzhuge/hadoop/hadoop-hdfs-project/hadoop-hdfs-native-client/target
 [exec] Start 1: test_test_libhdfs_threaded_hdfs_static
 [exec] 1/3 Test #1: test_test_libhdfs_threaded_hdfs_static ...   Passed
9.73 sec
 [exec] Start 2: test_test_libhdfs_zerocopy_hdfs_static
 [exec] 2/3 Test #2: test_test_libhdfs_zerocopy_hdfs_static ...***Failed
6.56 sec
 [exec] Start 3: test_test_native_mini_dfs
 [exec] Errors while running CTest
 [exec] 3/3 Test #3: test_test_native_mini_dfs    Passed
7.45 sec
 [exec]
 [exec] 67% tests passed, 1 tests failed out of 3
 [exec]
 [exec] Total Test time (real) =  23.74 sec
 [exec]
 [exec] The following tests FAILED:
 [exec]   2 - test_test_libhdfs_zerocopy_hdfs_static (Failed)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org