[jira] [Created] (HADOOP-14386) Make trunk work with Guava 11.0.2 again

2017-05-04 Thread Andrew Wang (JIRA)
Andrew Wang created HADOOP-14386:


 Summary: Make trunk work with Guava 11.0.2 again
 Key: HADOOP-14386
 URL: https://issues.apache.org/jira/browse/HADOOP-14386
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 3.0.0-alpha3
Reporter: Andrew Wang


As an alternative to reverting or shading HADOOP-10101 (the upgrade of Guava 
from 11.0.2 to 21.0), HADOOP-14380 makes the Guava version configurable. 
However, it still doesn't compile with Guava 11.0.2, since HADOOP-10101 chose 
to use the moved Guava classes rather than replacing them with alternatives.

This JIRA aims to make Hadoop compatible with Guava 11.0.2 as well as 21.0 by 
replacing usage of these moved Guava classes.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Re: About 2.7.4 Release

2017-05-04 Thread Konstantin Shvachko
Great Zhe. Let's monitor the build.

I marked all jiras I knew of for inclusion into 2.7.4 as I described before.
Target Version/s: 2.7.4
Label: release-blocker

Here is the link to the list: https://s.apache.org/Dzg4
Please let me know if I missed anything.
And feel free to pick up any. Most of backports are pretty straightforward,
but not all.

We can create tracking jiras for backporting if you need to run Jenkins on
the patch (and since Allen does not allow reopening them).
But I think the final patch should be attached to the original jira.
Otherwise history will be hard to follow.

Thanks,
--Konstantin

On Wed, May 3, 2017 at 4:53 PM, Zhe Zhang  wrote:

> Thanks for volunteering as RM Konstantin! The plan LGTM.
>
> I've created a nightly Jenkins job for branch-2.7 (unit tests):
> https://builds.apache.org/job/Hadoop-branch2.7-nightly/
>
> On Wed, May 3, 2017 at 12:42 AM Konstantin Shvachko 
> wrote:
>
>> Hey guys,
>>
>> I and a few of my colleagues would like to help here and move 2.7.4
>> release
>> forward. A few points in this regard.
>>
>> 1. Reading through this thread since March 1 I see that Vinod hinted on
>> managing the release. Vinod, if you still want the job / have bandwidth
>> will be happy to work with you.
>> Otherwise I am glad to volunteer as the release manager.
>>
>> 2. In addition to current blockers and criticals, I would like to propose
>> a
>> few issues to be included in the release, see the list below. Those are
>> mostly bug fixes and optimizations, which we already have in our internal
>> branch and run in production. Plus one minor feature "node labeling",
>> which
>> we found very handy, when you have heterogeneous environments and mixed
>> workloads, like MR and Spark.
>>
>> 3. For marking issues for the release I propose to
>>  - set the target version to 2.7.4, and
>>  - add a new label "release-blocker"
>> That way we will know issues targeted for the release without reopening
>> them for backports.
>>
>> 4. I see quite a few people are interested in the release. With all the
>> help I think we can target to release by the end of May.
>>
>> Other things include fixing CHANGES.txt and fixing Jenkins build for 2.7.4
>> branch.
>>
>> Thanks,
>> --Konstantin
>>
>> ==  List of issue for 2.7.4  ===
>> -- Backports
>> HADOOP-12975 . Add du
>> jitters
>> HDFS-9710 . IBR batching
>> HDFS-10715 . NPE when
>> applying AvailableSpaceBlockPlacementPolicy
>> HDFS-2538 . fsck removal
>> of dot printing
>> HDFS-8131 .
>> space-balanced
>> policy for balancer
>> HDFS-8549 . abort
>> balancer
>> if upgrade in progress
>> HDFS-9412 . skip small
>> blocks in getBlocks
>>
>> YARN-1471 . SLS
>> simulator
>> YARN-4302 . SLS
>> YARN-4367 . SLS
>> YARN-4612 . SLS
>>
>> - Node labeling
>> MAPREDUCE-6304 
>> YARN-2943 
>> YARN-4109 
>> YARN-4140 
>> YARN-4250 
>> YARN-4925 
>>
> --
> Zhe Zhang
> Apache Hadoop Committer
> http://zhe-thoughts.github.io/about/ | @oldcap
>


[jira] [Created] (HADOOP-14385) HttpExceptionUtils#validateResponse hides exceptions

2017-05-04 Thread Wei-Chiu Chuang (JIRA)
Wei-Chiu Chuang created HADOOP-14385:


 Summary: HttpExceptionUtils#validateResponse hides exceptions
 Key: HADOOP-14385
 URL: https://issues.apache.org/jira/browse/HADOOP-14385
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Wei-Chiu Chuang
Assignee: Wei-Chiu Chuang


In the following code
{code:title=HttpExceptionUtils#validateResponse}
try {
es = conn.getErrorStream();
ObjectMapper mapper = new ObjectMapper();
Map json = mapper.readValue(es, Map.class);
json = (Map) json.get(ERROR_JSON);
String exClass = (String) json.get(ERROR_CLASSNAME_JSON);
String exMsg = (String) json.get(ERROR_MESSAGE_JSON);
if (exClass != null) {
  try {
ClassLoader cl = HttpExceptionUtils.class.getClassLoader();
Class klass = cl.loadClass(exClass);
Constructor constr = klass.getConstructor(String.class);
toThrow = (Exception) constr.newInstance(exMsg);
  } catch (Exception ex) {
toThrow = new IOException(String.format(
"HTTP status [%d], exception [%s], message [%s] ",
conn.getResponseCode(), exClass, exMsg));
  }
} else {
  String msg = (exMsg != null) ? exMsg : conn.getResponseMessage();
  toThrow = new IOException(String.format(
  "HTTP status [%d], message [%s]", conn.getResponseCode(), msg));
}
  } catch (Exception ex) {
toThrow = new IOException(String.format( <-- here
"HTTP status [%d], message [%s]", conn.getResponseCode(),
conn.getResponseMessage()));
  }
{code}
If the an exception is thrown within the try block, the initial exception is 
swallowed, and it doesn't help debugging.
We had to cross reference this exception with the KMS server side to guess what 
happened.

IMHO the IOException thrown should also carry the initial exception. It should 
also print exClass and exMsg. It probably failed to instantiate an exception 
class.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14384) Reduce the visibility of {{FileSystem#newFSDataOutputStreamBuilder}} before the API becomes stable

2017-05-04 Thread Lei (Eddy) Xu (JIRA)
Lei (Eddy) Xu created HADOOP-14384:
--

 Summary: Reduce the visibility of 
{{FileSystem#newFSDataOutputStreamBuilder}} before the API becomes stable
 Key: HADOOP-14384
 URL: https://issues.apache.org/jira/browse/HADOOP-14384
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs
Affects Versions: 2.9.0
Reporter: Lei (Eddy) Xu
Assignee: Lei (Eddy) Xu
Priority: Minor


Before {{HADOOP-14365}} finishes, we should limit the this API within Hadoop 
project to prevent it being used by end users or other projects.





--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-14207) "dfsadmin -refreshCallQueue" fails with DecayRpcScheduler

2017-05-04 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14207?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HADOOP-14207.
-
Resolution: Fixed

I've pushed the fix to branch-2.8 and branch-2.8.1. Thanks!

> "dfsadmin -refreshCallQueue" fails with DecayRpcScheduler
> -
>
> Key: HADOOP-14207
> URL: https://issues.apache.org/jira/browse/HADOOP-14207
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: rpc-server
>Reporter: Surendra Singh Lilhore
>Assignee: Surendra Singh Lilhore
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: HADOOP-14207.001.patch, HADOOP-14207.002.patch, 
> HADOOP-14207.003.patch, HADOOP-14207.004.patch, HADOOP-14207.005.patch, 
> HADOOP-14207.006.patch
>
>
> {noformat}
> java.lang.RuntimeException: org.apache.hadoop.ipc.DecayRpcScheduler could not 
> be constructed.
> at 
> org.apache.hadoop.ipc.CallQueueManager.createScheduler(CallQueueManager.java:89)
> at 
> org.apache.hadoop.ipc.CallQueueManager.swapQueue(CallQueueManager.java:260)
> at org.apache.hadoop.ipc.Server.refreshCallQueue(Server.java:650)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.refreshCallQueue(NameNodeRpcServer.java:1582)
> at 
> org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolServerSideTranslatorPB.refreshCallQueue(RefreshCallQueueProtocolServerSideTranslatorPB.java:49)
> at 
> org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos$RefreshCallQueueProtocolService$2.callBlockingMethod(RefreshCallQueueProtocolProtos.java:769)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:845)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:788)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2455)
> Caused by: org.apache.hadoop.metrics2.MetricsException: Metrics source 
> DecayRpcSchedulerMetrics2.ipc.65110 already exists!
> at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.newSourceName(DefaultMetricsSystem.java:144)
> at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.sourceName(DefaultMetricsSystem.java:117)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/ppc64le

2017-05-04 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/

[May 3, 2017 2:47:17 PM] (kihwal) HADOOP-14372. TestSymlinkLocalFS timeouts are 
too low. Contributed by
[May 3, 2017 6:22:44 PM] (liuml07) HDFS-11739. Fix regression in tests caused 
by YARN-679. Contributed by
[May 3, 2017 9:49:30 PM] (sidharta) YARN-6374. Improve test coverage and add 
utility classes for common
[May 3, 2017 10:46:08 PM] (arp) HDFS-11488. JN log segment syncing should 
support HA upgrade.
[May 3, 2017 11:28:47 PM] (arp) Revert "HDFS-11488. JN log segment syncing 
should support HA upgrade.
[May 3, 2017 11:29:30 PM] (arp) HDFS-11722. Change Datanode file IO profiling 
sampling to percentage.




-1 overall


The following subsystems voted -1:
compile mvninstall unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc javac


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure190 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure120 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure140 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure180 
   hadoop.hdfs.server.namenode.TestFSImage 
   hadoop.hdfs.qjournal.server.TestJournalNode 
   hadoop.hdfs.TestDFSUpgrade 
   hadoop.hdfs.server.namenode.ha.TestBootstrapStandby 
   hadoop.hdfs.TestDistributedFileSystem 
   hadoop.hdfs.tools.offlineImageViewer.TestOfflineImageViewer 
   hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations 
   hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure100 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureReporting 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure110 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure010 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure090 
   hadoop.hdfs.TestDFSRSDefault10x4StripedOutputStreamWithFailure 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailureToleration 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure080 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure000 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.server.datanode.TestDataNodeUUID 
   hadoop.hdfs.server.datanode.fsdataset.impl.TestSpaceReservation 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure040 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure210 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure050 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure030 
   hadoop.mapreduce.v2.hs.TestHistoryServerLeveldbStateStoreService 
   hadoop.mapred.TestShuffleHandler 
   hadoop.tools.TestHadoopArchiveLogsRunner 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.yarn.server.timeline.TestRollingLevelDB 
   hadoop.yarn.server.timeline.TestTimelineDataManager 
   hadoop.yarn.server.timeline.TestLeveldbTimelineStore 
   hadoop.yarn.server.timeline.recovery.TestLeveldbTimelineStateStore 
   hadoop.yarn.server.timeline.TestRollingLevelDBTimelineStore 
   
hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer 
   hadoop.yarn.server.resourcemanager.TestRMEmbeddedElector 
   hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer 
   hadoop.yarn.server.resourcemanager.recovery.TestLeveldbRMStateStore 
   hadoop.yarn.server.TestDiskFailures 
   hadoop.yarn.server.TestMiniYarnClusterNodeUtilization 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.server.timeline.TestLevelDBCacheTimelineStore 
   hadoop.yarn.server.timeline.TestOverrideTimelineStoreYarnClient 
   hadoop.yarn.server.timeline.TestEntityGroupFSTimelineStore 

Timed out junit tests :

   org.apache.hadoop.hdfs.qjournal.client.TestQJMWithFaults 
   org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache 
   
org.apache.hadoop.yarn.server.resourcemanager.recovery.TestZKRMStateStorePerf 
  

   mvninstall:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/artifact/out/patch-mvninstall-root.txt
  [496K]

   compile:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/artifact/out/patch-compile-root.txt
  [20K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/artifact/out/patch-compile-root.txt
  [20K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/artifact/out/patch-compile-root.txt
  [20K]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-ppc/304/artifact/out/patch-unit-hadoop-assemblies.txt
  [4.0K]
   
https://build

[jira] [Reopened] (HADOOP-14207) "dfsadmin -refreshCallQueue" fails with DecayRpcScheduler

2017-05-04 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-14207?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du reopened HADOOP-14207:
-

> "dfsadmin -refreshCallQueue" fails with DecayRpcScheduler
> -
>
> Key: HADOOP-14207
> URL: https://issues.apache.org/jira/browse/HADOOP-14207
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: rpc-server
>Reporter: Surendra Singh Lilhore
>Assignee: Surendra Singh Lilhore
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: HADOOP-14207.001.patch, HADOOP-14207.002.patch, 
> HADOOP-14207.003.patch, HADOOP-14207.004.patch, HADOOP-14207.005.patch, 
> HADOOP-14207.006.patch
>
>
> {noformat}
> java.lang.RuntimeException: org.apache.hadoop.ipc.DecayRpcScheduler could not 
> be constructed.
> at 
> org.apache.hadoop.ipc.CallQueueManager.createScheduler(CallQueueManager.java:89)
> at 
> org.apache.hadoop.ipc.CallQueueManager.swapQueue(CallQueueManager.java:260)
> at org.apache.hadoop.ipc.Server.refreshCallQueue(Server.java:650)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.refreshCallQueue(NameNodeRpcServer.java:1582)
> at 
> org.apache.hadoop.ipc.protocolPB.RefreshCallQueueProtocolServerSideTranslatorPB.refreshCallQueue(RefreshCallQueueProtocolServerSideTranslatorPB.java:49)
> at 
> org.apache.hadoop.ipc.proto.RefreshCallQueueProtocolProtos$RefreshCallQueueProtocolService$2.callBlockingMethod(RefreshCallQueueProtocolProtos.java:769)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:845)
> at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:788)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1807)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2455)
> Caused by: org.apache.hadoop.metrics2.MetricsException: Metrics source 
> DecayRpcSchedulerMetrics2.ipc.65110 already exists!
> at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.newSourceName(DefaultMetricsSystem.java:144)
> at 
> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.sourceName(DefaultMetricsSystem.java:117)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14382) Remove usages of MoreObjects.toStringHelper

2017-05-04 Thread Andrew Wang (JIRA)
Andrew Wang created HADOOP-14382:


 Summary: Remove usages of MoreObjects.toStringHelper
 Key: HADOOP-14382
 URL: https://issues.apache.org/jira/browse/HADOOP-14382
 Project: Hadoop Common
  Issue Type: Improvement
  Components: metrics
Affects Versions: 3.0.0-alpha2
Reporter: Andrew Wang
Assignee: Andrew Wang
Priority: Minor


MoreObjects.toStringHelper is a source of incompatibility across Guava 
versions. Let's move off of this to a native Java 8 API.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-14381) S3AUtils.translateException to map 503 reponse to => throttling failure

2017-05-04 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-14381:
---

 Summary: S3AUtils.translateException to map 503 reponse to => 
throttling failure
 Key: HADOOP-14381
 URL: https://issues.apache.org/jira/browse/HADOOP-14381
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 2.8.0
Reporter: Steve Loughran


When AWS S3 returns "503", it means that the overall set of requests on a part 
of an S3 bucket exceeds the permitted limit; the client(s) need to throttle 
back or away for some rebalancing to complete.

The aws SDK retries 3 times on a 503, but then throws it up. Our code doesn't 
do anything with that other than create a generic {{AWSS3IOException}}.

Proposed
* add a new exception, {{AWSOverloadedException}}
* raise it on a 503 from S3 (& for s3guard, on DDB complaints)
* have it include a link to a wiki page on the topic, as well as the path
* and any other diags

Code talking to S3 may then be able to catch this and choose to react. Some 
retry with exponential backoff is the obvious option. Failing, well, that could 
trigger task reattempts at that part of the query, then job retry —which will 
again fail, *unless the number of tasks run in parallel is reduced*

As this throttling is across all clients talking to the same part of a bucket, 
fixing it is potentially a high level option. We can at least start by 
reporting things better




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2017-05-04 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/393/

[May 3, 2017 2:47:17 PM] (kihwal) HADOOP-14372. TestSymlinkLocalFS timeouts are 
too low. Contributed by
[May 3, 2017 6:22:44 PM] (liuml07) HDFS-11739. Fix regression in tests caused 
by YARN-679. Contributed by
[May 3, 2017 9:49:30 PM] (sidharta) YARN-6374. Improve test coverage and add 
utility classes for common
[May 3, 2017 10:46:08 PM] (arp) HDFS-11488. JN log segment syncing should 
support HA upgrade.
[May 3, 2017 11:28:47 PM] (arp) Revert "HDFS-11488. JN log segment syncing 
should support HA upgrade.
[May 3, 2017 11:29:30 PM] (arp) HDFS-11722. Change Datanode file IO profiling 
sampling to percentage.




-1 overall


The following subsystems voted -1:
findbugs unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

FindBugs :

   module:hadoop-common-project/hadoop-minikdc 
   Possible null pointer dereference in 
org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value of called 
method Dereferenced at 
MiniKdc.java:org.apache.hadoop.minikdc.MiniKdc.delete(File) due to return value 
of called method Dereferenced at MiniKdc.java:[line 368] 

FindBugs :

   module:hadoop-common-project/hadoop-auth 
   
org.apache.hadoop.security.authentication.server.MultiSchemeAuthenticationHandler.authenticate(HttpServletRequest,
 HttpServletResponse) makes inefficient use of keySet iterator instead of 
entrySet iterator At MultiSchemeAuthenticationHandler.java:of keySet iterator 
instead of entrySet iterator At MultiSchemeAuthenticationHandler.java:[line 
192] 

FindBugs :

   module:hadoop-common-project/hadoop-common 
   org.apache.hadoop.crypto.CipherSuite.setUnknownValue(int) 
unconditionally sets the field unknownValue At CipherSuite.java:unknownValue At 
CipherSuite.java:[line 44] 
   org.apache.hadoop.crypto.CryptoProtocolVersion.setUnknownValue(int) 
unconditionally sets the field unknownValue At 
CryptoProtocolVersion.java:unknownValue At CryptoProtocolVersion.java:[line 67] 
   Possible null pointer dereference in 
org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to return value of 
called method Dereferenced at 
FileUtil.java:org.apache.hadoop.fs.FileUtil.fullyDeleteOnExit(File) due to 
return value of called method Dereferenced at FileUtil.java:[line 118] 
   Possible null pointer dereference in 
org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path, 
File, Path, File) due to return value of called method Dereferenced at 
RawLocalFileSystem.java:org.apache.hadoop.fs.RawLocalFileSystem.handleEmptyDstDirectoryOnWindows(Path,
 File, Path, File) due to return value of called method Dereferenced at 
RawLocalFileSystem.java:[line 387] 
   Return value of org.apache.hadoop.fs.permission.FsAction.or(FsAction) 
ignored, but method has no side effect At FTPFileSystem.java:but method has no 
side effect At FTPFileSystem.java:[line 421] 
   Useless condition:lazyPersist == true at this point At 
CommandWithDestination.java:[line 502] 
   org.apache.hadoop.io.DoubleWritable.compareTo(DoubleWritable) 
incorrectly handles double value At DoubleWritable.java: At 
DoubleWritable.java:[line 78] 
   org.apache.hadoop.io.DoubleWritable$Comparator.compare(byte[], int, int, 
byte[], int, int) incorrectly handles double value At DoubleWritable.java:int) 
incorrectly handles double value At DoubleWritable.java:[line 97] 
   org.apache.hadoop.io.FloatWritable.compareTo(FloatWritable) incorrectly 
handles float value At FloatWritable.java: At FloatWritable.java:[line 71] 
   org.apache.hadoop.io.FloatWritable$Comparator.compare(byte[], int, int, 
byte[], int, int) incorrectly handles float value At FloatWritable.java:int) 
incorrectly handles float value At FloatWritable.java:[line 89] 
   Possible null pointer dereference in 
org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) due to return 
value of called method Dereferenced at 
IOUtils.java:org.apache.hadoop.io.IOUtils.listDirectory(File, FilenameFilter) 
due to return value of called method Dereferenced at IOUtils.java:[line 350] 
   org.apache.hadoop.io.erasurecode.ECSchema.toString() makes inefficient 
use of keySet iterator instead of entrySet iterator At ECSchema.java:keySet 
iterator instead of entrySet iterator At ECSchema.java:[line 193] 
   Possible bad parsing of shift operation in 
org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At 
Utils.java:operation in 
org.apache.hadoop.io.file.tfile.Utils$Version.hashCode() At Utils.java:[line 
398] 
   
org.apache.hadoop.metrics2.lib.DefaultMetricsFactory.setInstance(MutableMetricsFactory)
 unconditionally sets the field mmfImpl At DefaultMetricsFactory.java:mmfI

[jira] [Created] (HADOOP-14380) Make Guava version Hadoop builds with configurable

2017-05-04 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-14380:
---

 Summary: Make Guava version Hadoop builds with configurable
 Key: HADOOP-14380
 URL: https://issues.apache.org/jira/browse/HADOOP-14380
 Project: Hadoop Common
  Issue Type: Bug
  Components: build
Affects Versions: 3.0.0-alpha3
Reporter: Steve Loughran
Assignee: Steve Loughran


Make the choice of guava version Hadoop builds with configurable, so people 
building Hadoop 3 alphas can build with an older version and so cause less 
unhappiness downstream



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org