Build failed in Jenkins: Hadoop-Common-trunk #1527

2015-06-15 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Common-trunk/1527/changes

Changes:

[yzhang] HDFS-8596. TestDistributedFileSystem et al tests are broken in 
branch-2 due to incorrect setting of datanode attribute. Contributed by 
Yongjun Zhang.

[arp] HDFS-8595. TestCommitBlockSynchronization fails in branch-2.7. (Patch 
applies to all branches). (Contributed by Arpit Agarwal)

--
[...truncated 5173 lines...]
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.053 sec - in 
org.apache.hadoop.util.TestFindClass
Running org.apache.hadoop.util.TestLightWeightGSet
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.206 sec - in 
org.apache.hadoop.util.TestLightWeightGSet
Running org.apache.hadoop.util.TestPureJavaCrc32
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.703 sec - in 
org.apache.hadoop.util.TestPureJavaCrc32
Running org.apache.hadoop.util.TestStringUtils
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.461 sec - in 
org.apache.hadoop.util.TestStringUtils
Running org.apache.hadoop.util.TestProtoUtil
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.482 sec - in 
org.apache.hadoop.util.TestProtoUtil
Running org.apache.hadoop.util.TestSignalLogger
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.279 sec - in 
org.apache.hadoop.util.TestSignalLogger
Running org.apache.hadoop.util.TestDiskChecker
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.874 sec - in 
org.apache.hadoop.util.TestDiskChecker
Running org.apache.hadoop.util.TestShutdownThreadsHelper
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.324 sec - in 
org.apache.hadoop.util.TestShutdownThreadsHelper
Running org.apache.hadoop.util.TestCacheableIPList
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.421 sec - in 
org.apache.hadoop.util.TestCacheableIPList
Running org.apache.hadoop.util.TestLineReader
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.349 sec - in 
org.apache.hadoop.util.TestLineReader
Running org.apache.hadoop.util.TestIdentityHashStore
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.28 sec - in 
org.apache.hadoop.util.TestIdentityHashStore
Running org.apache.hadoop.util.TestClasspath
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.676 sec - in 
org.apache.hadoop.util.TestClasspath
Running org.apache.hadoop.util.TestApplicationClassLoader
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.329 sec - in 
org.apache.hadoop.util.TestApplicationClassLoader
Running org.apache.hadoop.util.TestShell
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.573 sec - in 
org.apache.hadoop.util.TestShell
Running org.apache.hadoop.util.TestShutdownHookManager
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.436 sec - in 
org.apache.hadoop.util.TestShutdownHookManager
Running org.apache.hadoop.util.TestConfTest
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.613 sec - in 
org.apache.hadoop.util.TestConfTest
Running org.apache.hadoop.util.TestHttpExceptionUtils
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.963 sec - in 
org.apache.hadoop.util.TestHttpExceptionUtils
Running org.apache.hadoop.util.TestJarFinder
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.863 sec - in 
org.apache.hadoop.util.TestJarFinder
Running org.apache.hadoop.util.hash.TestHash
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.464 sec - in 
org.apache.hadoop.util.hash.TestHash
Running org.apache.hadoop.util.TestLightWeightCache
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.486 sec - in 
org.apache.hadoop.util.TestLightWeightCache
Running org.apache.hadoop.util.TestNativeCodeLoader
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.174 sec - in 
org.apache.hadoop.util.TestNativeCodeLoader
Running org.apache.hadoop.util.TestReflectionUtils
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.698 sec - in 
org.apache.hadoop.util.TestReflectionUtils
Running org.apache.hadoop.crypto.TestCryptoStreams
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.851 sec - 
in org.apache.hadoop.crypto.TestCryptoStreams
Running org.apache.hadoop.crypto.TestCryptoStreamsForLocalFS
Tests run: 14, Failures: 0, Errors: 0, Skipped: 5, Time elapsed: 17.942 sec - 
in org.apache.hadoop.crypto.TestCryptoStreamsForLocalFS
Running org.apache.hadoop.crypto.TestOpensslCipher
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.3 sec - in 
org.apache.hadoop.crypto.TestOpensslCipher
Running org.apache.hadoop.crypto.TestCryptoStreamsWithOpensslAesCtrCryptoCodec
Tests run: 14, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.592 sec - 
in org.apache.hadoop.crypto.TestCryptoStreamsWithOpensslAesCtrCryptoCodec
Running org.apache.hadoop.crypto.TestCryptoStreamsNormal
Tests run: 

Build failed in Jenkins: Hadoop-common-trunk-Java8 #229

2015-06-15 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-common-trunk-Java8/229/changes

Changes:

[yzhang] HDFS-8596. TestDistributedFileSystem et al tests are broken in 
branch-2 due to incorrect setting of datanode attribute. Contributed by 
Yongjun Zhang.

[arp] HDFS-8595. TestCommitBlockSynchronization fails in branch-2.7. (Patch 
applies to all branches). (Contributed by Arpit Agarwal)

--
[...truncated 5535 lines...]
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.707 sec - in 
org.apache.hadoop.security.TestAuthenticationFilter
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestGroupFallback
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.798 sec - in 
org.apache.hadoop.security.TestGroupFallback
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestProxyUserFromEnv
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.684 sec - in 
org.apache.hadoop.security.TestProxyUserFromEnv
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestUGILoginFromKeytab
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.404 sec - in 
org.apache.hadoop.security.TestUGILoginFromKeytab
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestLdapGroupsMappingWithPosixGroup
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.672 sec - in 
org.apache.hadoop.security.TestLdapGroupsMappingWithPosixGroup
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestShellBasedIdMapping
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.862 sec - in 
org.apache.hadoop.security.TestShellBasedIdMapping
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestDoAsEffectiveUser
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.333 sec - in 
org.apache.hadoop.security.TestDoAsEffectiveUser
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestUGIWithExternalKdc
Tests run: 1, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 0.076 sec - in 
org.apache.hadoop.security.TestUGIWithExternalKdc
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestJNIGroupsMapping
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.469 sec - in 
org.apache.hadoop.security.TestJNIGroupsMapping
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestGroupsCaching
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.643 sec - in 
org.apache.hadoop.security.TestGroupsCaching
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.alias.TestCredentialProviderFactory
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.179 sec - in 
org.apache.hadoop.security.alias.TestCredentialProviderFactory
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.alias.TestCredentialProvider
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.1 sec - in 
org.apache.hadoop.security.alias.TestCredentialProvider
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.alias.TestCredShell
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.332 sec - in 
org.apache.hadoop.security.alias.TestCredShell
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.ssl.TestReloadingX509TrustManager
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.61 sec - in 
org.apache.hadoop.security.ssl.TestReloadingX509TrustManager
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.ssl.TestSSLFactory
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.389 sec - in 
org.apache.hadoop.security.ssl.TestSSLFactory
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; 
support was removed in 8.0
Running org.apache.hadoop.security.TestLdapGroupsMapping
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.246 sec - in 

Re: set up jenkins test for branch-2

2015-06-15 Thread Allen Wittenauer

On Jun 14, 2015, at 1:00 PM, Yongjun Zhang yzh...@cloudera.com wrote:

 From time to time we saw changes that work fine in trunk but not branch-2,
 and we don't catch the issue in a timely manner. The difference between
 trunk and branch-2 is sufficient to justify periodic jenkins test and even
 pre-commit test for branch-2.

What is your expected outcome?

The current nightly builds have been failing pretty consistently for a 
while now.  Adding another one isn't going to make the situation any better 
since everyone pretty much ignores them.  As I said at the HDFS BOF:  we 
desperately need a moratorium on adding new features and need to work on 
stabilizing both branch-2 and trunk.  We seem to be in a quantity over 
quality mode and it's slowly killing the project.  We haven't had a stable 
release since 2.4.1, which is approaching the 1 year mark.

(Yes, I know.  Various vendors have shipped releases since 2.4.1 as 
part of their distributions.  Before saying anything, perhaps you should spend 
some time with your support folks….)

As Sean pointed out, precommit for branch-2 has been working for a 
while now if one provides enough hints to test-patch as part of the patch name.

Re: Protocol Buffers version

2015-06-15 Thread Andrew Purtell
I can't answer the original question but can point out the protostuff (
https://github.com/protostuff/protostuff) folks have been responsive and
friendly in the past when we (HBase) were curious about swapping in their
stuff. Two significant benefits of protostuff, IMHO, is ASL 2 licensing and
everything is implemented in Java including the compiler.


On Mon, Jun 15, 2015 at 8:49 AM, Sean Busbey bus...@cloudera.com wrote:

 Anyone have a read on how the protobuf folks would feel about that? Apache
 has a history of not accepting projects that are non-amicable forks.

 On Mon, Jun 15, 2015 at 9:24 AM, Allen Wittenauer a...@altiscale.com
 wrote:

 
  On Jun 12, 2015, at 1:03 PM, Alan Burlison alan.burli...@oracle.com
  wrote:
 
   On 14/05/2015 18:41, Chris Nauroth wrote:
  
   As a reminder though, the community probably would want to see a
 strong
   justification for the upgrade in terms of features or performance or
   something else.  Right now, I'm not seeing a significant benefit for
 us
   based on my reading of their release notes.  I think it's worthwhile
 to
   figure this out first.  Otherwise, there is a risk that any testing
 work
   turns out to be a wasted effort.
  
   One reason at least: PB 2.5.0 has no support for Solaris SPARC. 2.6.1
  does.
 
 
  That's a pretty good reason.
 
  Some of us had a discussion at Summit about effectively forking
  protobuf and making it an Apache TLP.  This would give us a chance to get
  out from under Google's blind spot, guarantee better compatibility across
  the ecosystem, etc, etc.
 
  It is sounding more and more like that's really what needs to
  happen.




 --
 Sean




-- 
Best regards,

   - Andy

Problems worthy of attack prove their worth by hitting back. - Piet Hein
(via Tom White)


Re: Protocol Buffers version

2015-06-15 Thread Allen Wittenauer

On Jun 12, 2015, at 1:03 PM, Alan Burlison alan.burli...@oracle.com wrote:

 On 14/05/2015 18:41, Chris Nauroth wrote:
 
 As a reminder though, the community probably would want to see a strong
 justification for the upgrade in terms of features or performance or
 something else.  Right now, I'm not seeing a significant benefit for us
 based on my reading of their release notes.  I think it's worthwhile to
 figure this out first.  Otherwise, there is a risk that any testing work
 turns out to be a wasted effort.
 
 One reason at least: PB 2.5.0 has no support for Solaris SPARC. 2.6.1 does.


That's a pretty good reason.

Some of us had a discussion at Summit about effectively forking 
protobuf and making it an Apache TLP.  This would give us a chance to get out 
from under Google's blind spot, guarantee better compatibility across the 
ecosystem, etc, etc.

It is sounding more and more like that's really what needs to happen.

Re: Protocol Buffers version

2015-06-15 Thread Sean Busbey
Anyone have a read on how the protobuf folks would feel about that? Apache
has a history of not accepting projects that are non-amicable forks.

On Mon, Jun 15, 2015 at 9:24 AM, Allen Wittenauer a...@altiscale.com wrote:


 On Jun 12, 2015, at 1:03 PM, Alan Burlison alan.burli...@oracle.com
 wrote:

  On 14/05/2015 18:41, Chris Nauroth wrote:
 
  As a reminder though, the community probably would want to see a strong
  justification for the upgrade in terms of features or performance or
  something else.  Right now, I'm not seeing a significant benefit for us
  based on my reading of their release notes.  I think it's worthwhile to
  figure this out first.  Otherwise, there is a risk that any testing work
  turns out to be a wasted effort.
 
  One reason at least: PB 2.5.0 has no support for Solaris SPARC. 2.6.1
 does.


 That's a pretty good reason.

 Some of us had a discussion at Summit about effectively forking
 protobuf and making it an Apache TLP.  This would give us a chance to get
 out from under Google's blind spot, guarantee better compatibility across
 the ecosystem, etc, etc.

 It is sounding more and more like that's really what needs to
 happen.




-- 
Sean


Re: [DISCUSS] project for pre-commit patch testing (was Re: upstream jenkins build broken?)

2015-06-15 Thread Josh Elser

+1

(Have been talking to Sean in private on the subject -- seems 
appropriate to voice some public support)


I'd be interested in this for Accumulo and Slider. For Accumulo, we've 
come a far way without a pre-commit build, primarily due to a CTR 
process. We have seen the repeated questions of how do I run the tests 
which a more automated workflow would help with, IMO. I think Slider 
could benefit with the same reasons.


I'd also be giddy to see the recent improvements in Hadoop trickle down 
into the other projects that Allen already mentioned.


Take this as record that I'd be happy to try to help out where possible.

Sean Busbey wrote:

thank you for making a more digestible version Allen. :)

If you're interested in soliciting feedback from other projects, I created
ASF short links to this thread in common-dev and hbase:


* http://s.apache.org/yetus-discuss-hadoop
* http://s.apache.org/yetus-discuss-hbase

While I agree that it's important to get feedback from ASF projects that
might find this useful, I can say that recently I've been involved in the
non-ASF project YCSB and both the pretest and better shell stuff would be
immensely useful over there.

On Mon, Jun 15, 2015 at 10:36 PM, Allen Wittenauera...@altiscale.com  wrote:


 I'm clearly +1 on this idea.  As part of the rewrite in Hadoop of
test-patch, it was amazing to see how far and wide this bit of code as
spread.  So I see consolidating everyone's efforts as a huge win for a
large number of projects.  (esp considering how many I saw suffering from a
variety of identified bugs! )

 But….

 I think it's important for people involved in those other projects
to speak up and voice an opinion as to whether this is useful.

To summarize:

 In the short term, a single location to get/use a precommit patch
tester rather than everyone building/supporting their own in their spare
time.

  FWIW, we've already got the code base modified to be pluggable.
We've written some basic/simple plugins that support Hadoop, HBase, Tajo,
Tez, Pig, and Flink.  For HBase and Flink, this does include their custom
checks.  Adding support for other project shouldn't be hard.  Simple
projects take almost no time after seeing the basic pattern.

 I think it's worthwhile highlighting that means support for both
JIRA and GitHub as well as Ant and Maven from the same code base.

Longer term:

 Well, we clearly have ideas of things that we want to do. Adding
more features to test-patch (review board? gradle?) is obvious. But what
about teasing apart and generalizing some of the other shell bits from
projects? A common library for building CLI tools to fault injection to
release documentation creation tools to …  I'd even like to see us get as
advanced as a run this program to auto-generate daemon stop/start bits.

 I had a few chats with people about this idea at Hadoop Summit.
What's truly exciting are the ideas that people had once they realized what
kinds of problems we're trying to solve.  It's always amazing the problems
that projects have that could be solved by these types of solutions.  Let's
stop hiding our cool toys in this area.

 So, what feedback and ideas do you have in this area?  Are you a
yay or a nay?


On Jun 15, 2015, at 4:47 PM, Sean Busbeybus...@cloudera.com  wrote:


Oof. I had meant to push on this again but life got in the way and now

the

June board meeting is upon us. Sorry everyone. In the event that this

ends

up contentious, hopefully one of the copied communities can give us a
branch to work in.

I know everyone is busy, so here's the short version of this email: I'd
like to move some of the code currently in Hadoop (test-patch) into a new
TLP focused on QA tooling. I'm not sure what the best format for priming
this conversation is. ORC filled in the incubator project proposal
template, but I'm not sure how much that confused the issue. So to start,
I'll just write what I'm hoping we can accomplish in general terms here.

All software development projects that are community based (that is,
accepting outside contributions) face a common QA problem for vetting
in-coming contributions. Hadoop is fortunate enough to be sufficiently
popular that the weight of the problem drove tool development (i.e.
test-patch). That tool is generalizable enough that a bunch of other TLPs
have adopted their own forks. Unfortunately, in most projects this kind

of

QA work is an enabler rather than a primary concern, so often the tooling
is worked on ad-hoc and little shared improvements happen across
projects. Since
the tooling itself is never a primary concern, any made is rarely reused
outside of ASF projects.

Over the last couple months a few of us have been working on generalizing
the tooling present in the Hadoop code base (because it was the most

mature

out of all those in the various projects) and it's reached a point where

we

think we can start bringing on other downstream 

Re: [DISCUSS] project for pre-commit patch testing (was Re: upstream jenkins build broken?)

2015-06-15 Thread Chris Nauroth
+1

ZooKeeper is another project that has expressed interest in improving its
pre-commit process lately.  I understand Allen has had some success
applying this to the ZooKeeper build too, with some small caveats around
quirks in the build.xml that I think we can resolve.

I'm interested in defining how the release model works for a project like
this.  The current model of forking and checking it in directly to
multiple projects leads to the fragmentation and bugs described earlier in
the thread.  Another possible model is something more dynamic, like a
bootstrap script capable of checking out a release from a git tag before
launching pre-commit.  I'm interested to hear from various projects on how
they'd like to integrate.

--Chris Nauroth




On 6/15/15, 8:57 PM, Josh Elser els...@apache.org wrote:

+1

(Have been talking to Sean in private on the subject -- seems
appropriate to voice some public support)

I'd be interested in this for Accumulo and Slider. For Accumulo, we've
come a far way without a pre-commit build, primarily due to a CTR
process. We have seen the repeated questions of how do I run the tests
which a more automated workflow would help with, IMO. I think Slider
could benefit with the same reasons.

I'd also be giddy to see the recent improvements in Hadoop trickle down
into the other projects that Allen already mentioned.

Take this as record that I'd be happy to try to help out where possible.

Sean Busbey wrote:
 thank you for making a more digestible version Allen. :)

 If you're interested in soliciting feedback from other projects, I
created
 ASF short links to this thread in common-dev and hbase:


 * http://s.apache.org/yetus-discuss-hadoop
 * http://s.apache.org/yetus-discuss-hbase

 While I agree that it's important to get feedback from ASF projects that
 might find this useful, I can say that recently I've been involved in
the
 non-ASF project YCSB and both the pretest and better shell stuff would
be
 immensely useful over there.

 On Mon, Jun 15, 2015 at 10:36 PM, Allen Wittenauera...@altiscale.com
wrote:

  I'm clearly +1 on this idea.  As part of the rewrite in
Hadoop of
 test-patch, it was amazing to see how far and wide this bit of code as
 spread.  So I see consolidating everyone's efforts as a huge win for a
 large number of projects.  (esp considering how many I saw suffering
from a
 variety of identified bugs! )

  But….

  I think it's important for people involved in those other
projects
 to speak up and voice an opinion as to whether this is useful.

 To summarize:

  In the short term, a single location to get/use a precommit
patch
 tester rather than everyone building/supporting their own in their
spare
 time.

   FWIW, we've already got the code base modified to be
pluggable.
 We've written some basic/simple plugins that support Hadoop, HBase,
Tajo,
 Tez, Pig, and Flink.  For HBase and Flink, this does include their
custom
 checks.  Adding support for other project shouldn't be hard.  Simple
 projects take almost no time after seeing the basic pattern.

  I think it's worthwhile highlighting that means support for
both
 JIRA and GitHub as well as Ant and Maven from the same code base.

 Longer term:

  Well, we clearly have ideas of things that we want to do.
Adding
 more features to test-patch (review board? gradle?) is obvious. But
what
 about teasing apart and generalizing some of the other shell bits from
 projects? A common library for building CLI tools to fault injection to
 release documentation creation tools to …  I'd even like to see us get
as
 advanced as a run this program to auto-generate daemon stop/start
bits.

  I had a few chats with people about this idea at Hadoop
Summit.
 What's truly exciting are the ideas that people had once they realized
what
 kinds of problems we're trying to solve.  It's always amazing the
problems
 that projects have that could be solved by these types of solutions.
Let's
 stop hiding our cool toys in this area.

  So, what feedback and ideas do you have in this area?  Are
you a
 yay or a nay?


 On Jun 15, 2015, at 4:47 PM, Sean Busbeybus...@cloudera.com  wrote:

 Oof. I had meant to push on this again but life got in the way and now
 the
 June board meeting is upon us. Sorry everyone. In the event that this
 ends
 up contentious, hopefully one of the copied communities can give us a
 branch to work in.

 I know everyone is busy, so here's the short version of this email:
I'd
 like to move some of the code currently in Hadoop (test-patch) into a
new
 TLP focused on QA tooling. I'm not sure what the best format for
priming
 this conversation is. ORC filled in the incubator project proposal
 template, but I'm not sure how much that confused the issue. So to
start,
 I'll just write what I'm hoping we can accomplish in general terms
here.

 All software development projects that are community based (that is,
 accepting outside contributions) face a 

Re: [DISCUSS] project for pre-commit patch testing (was Re: upstream jenkins build broken?)

2015-06-15 Thread Allen Wittenauer

I'm clearly +1 on this idea.  As part of the rewrite in Hadoop of 
test-patch, it was amazing to see how far and wide this bit of code as spread.  
So I see consolidating everyone's efforts as a huge win for a large number of 
projects.  (esp considering how many I saw suffering from a variety of 
identified bugs! )

But….

I think it's important for people involved in those other projects to 
speak up and voice an opinion as to whether this is useful. 

To summarize:

In the short term, a single location to get/use a precommit patch 
tester rather than everyone building/supporting their own in their spare time. 

 FWIW, we've already got the code base modified to be pluggable.  We've 
written some basic/simple plugins that support Hadoop, HBase, Tajo, Tez, Pig, 
and Flink.  For HBase and Flink, this does include their custom checks.  Adding 
support for other project shouldn't be hard.  Simple projects take almost no 
time after seeing the basic pattern.

I think it's worthwhile highlighting that means support for both JIRA 
and GitHub as well as Ant and Maven from the same code base.

Longer term:

Well, we clearly have ideas of things that we want to do. Adding more 
features to test-patch (review board? gradle?) is obvious. But what about 
teasing apart and generalizing some of the other shell bits from projects? A 
common library for building CLI tools to fault injection to release 
documentation creation tools to …  I'd even like to see us get as advanced as a 
run this program to auto-generate daemon stop/start bits.

I had a few chats with people about this idea at Hadoop Summit.  What's 
truly exciting are the ideas that people had once they realized what kinds of 
problems we're trying to solve.  It's always amazing the problems that projects 
have that could be solved by these types of solutions.  Let's stop hiding our 
cool toys in this area.

So, what feedback and ideas do you have in this area?  Are you a yay or 
a nay?


On Jun 15, 2015, at 4:47 PM, Sean Busbey bus...@cloudera.com wrote:

 Oof. I had meant to push on this again but life got in the way and now the
 June board meeting is upon us. Sorry everyone. In the event that this ends
 up contentious, hopefully one of the copied communities can give us a
 branch to work in.
 
 I know everyone is busy, so here's the short version of this email: I'd
 like to move some of the code currently in Hadoop (test-patch) into a new
 TLP focused on QA tooling. I'm not sure what the best format for priming
 this conversation is. ORC filled in the incubator project proposal
 template, but I'm not sure how much that confused the issue. So to start,
 I'll just write what I'm hoping we can accomplish in general terms here.
 
 All software development projects that are community based (that is,
 accepting outside contributions) face a common QA problem for vetting
 in-coming contributions. Hadoop is fortunate enough to be sufficiently
 popular that the weight of the problem drove tool development (i.e.
 test-patch). That tool is generalizable enough that a bunch of other TLPs
 have adopted their own forks. Unfortunately, in most projects this kind of
 QA work is an enabler rather than a primary concern, so often the tooling
 is worked on ad-hoc and little shared improvements happen across
 projects. Since
 the tooling itself is never a primary concern, any made is rarely reused
 outside of ASF projects.
 
 Over the last couple months a few of us have been working on generalizing
 the tooling present in the Hadoop code base (because it was the most mature
 out of all those in the various projects) and it's reached a point where we
 think we can start bringing on other downstream users. This means we need
 to start establishing things like a release cadence and to grow the new
 contributors we have to handle more project responsibility. Personally, I
 think that means it's time to move out from under Hadoop to drive things as
 our own community. Eventually, I hope the community can help draw in a
 group of folks traditionally underrepresented in ASF projects, namely QA
 and operations folks.
 
 I think test-patch by itself has enough scope to justify a project. Having
 a solid set of build tools that are customizable to fit the norms of
 different software communities is a bunch of work. Making it work well in
 both the context of automated test systems like Jenkins and for individual
 developers is even more work. We could easily also take over maintenance of
 things like shelldocs, since test-patch is the primary consumer of that
 currently but it's generally useful tooling.
 
 In addition to test-patch, I think the proposed project has some future
 growth potential. Given some adoption of test-patch to prove utility, the
 project could build on the ties it makes to start building tools to help
 projects do their own longer-run testing. Note that I'm talking about the
 tools to build 

Re: [DISCUSS] project for pre-commit patch testing (was Re: upstream jenkins build broken?)

2015-06-15 Thread Sean Busbey
thank you for making a more digestible version Allen. :)

If you're interested in soliciting feedback from other projects, I created
ASF short links to this thread in common-dev and hbase:


* http://s.apache.org/yetus-discuss-hadoop
* http://s.apache.org/yetus-discuss-hbase

While I agree that it's important to get feedback from ASF projects that
might find this useful, I can say that recently I've been involved in the
non-ASF project YCSB and both the pretest and better shell stuff would be
immensely useful over there.

On Mon, Jun 15, 2015 at 10:36 PM, Allen Wittenauer a...@altiscale.com wrote:


 I'm clearly +1 on this idea.  As part of the rewrite in Hadoop of
 test-patch, it was amazing to see how far and wide this bit of code as
 spread.  So I see consolidating everyone's efforts as a huge win for a
 large number of projects.  (esp considering how many I saw suffering from a
 variety of identified bugs! )

 But….

 I think it's important for people involved in those other projects
 to speak up and voice an opinion as to whether this is useful.

 To summarize:

 In the short term, a single location to get/use a precommit patch
 tester rather than everyone building/supporting their own in their spare
 time.

  FWIW, we've already got the code base modified to be pluggable.
 We've written some basic/simple plugins that support Hadoop, HBase, Tajo,
 Tez, Pig, and Flink.  For HBase and Flink, this does include their custom
 checks.  Adding support for other project shouldn't be hard.  Simple
 projects take almost no time after seeing the basic pattern.

 I think it's worthwhile highlighting that means support for both
 JIRA and GitHub as well as Ant and Maven from the same code base.

 Longer term:

 Well, we clearly have ideas of things that we want to do. Adding
 more features to test-patch (review board? gradle?) is obvious. But what
 about teasing apart and generalizing some of the other shell bits from
 projects? A common library for building CLI tools to fault injection to
 release documentation creation tools to …  I'd even like to see us get as
 advanced as a run this program to auto-generate daemon stop/start bits.

 I had a few chats with people about this idea at Hadoop Summit.
 What's truly exciting are the ideas that people had once they realized what
 kinds of problems we're trying to solve.  It's always amazing the problems
 that projects have that could be solved by these types of solutions.  Let's
 stop hiding our cool toys in this area.

 So, what feedback and ideas do you have in this area?  Are you a
 yay or a nay?


 On Jun 15, 2015, at 4:47 PM, Sean Busbey bus...@cloudera.com wrote:

  Oof. I had meant to push on this again but life got in the way and now
 the
  June board meeting is upon us. Sorry everyone. In the event that this
 ends
  up contentious, hopefully one of the copied communities can give us a
  branch to work in.
 
  I know everyone is busy, so here's the short version of this email: I'd
  like to move some of the code currently in Hadoop (test-patch) into a new
  TLP focused on QA tooling. I'm not sure what the best format for priming
  this conversation is. ORC filled in the incubator project proposal
  template, but I'm not sure how much that confused the issue. So to start,
  I'll just write what I'm hoping we can accomplish in general terms here.
 
  All software development projects that are community based (that is,
  accepting outside contributions) face a common QA problem for vetting
  in-coming contributions. Hadoop is fortunate enough to be sufficiently
  popular that the weight of the problem drove tool development (i.e.
  test-patch). That tool is generalizable enough that a bunch of other TLPs
  have adopted their own forks. Unfortunately, in most projects this kind
 of
  QA work is an enabler rather than a primary concern, so often the tooling
  is worked on ad-hoc and little shared improvements happen across
  projects. Since
  the tooling itself is never a primary concern, any made is rarely reused
  outside of ASF projects.
 
  Over the last couple months a few of us have been working on generalizing
  the tooling present in the Hadoop code base (because it was the most
 mature
  out of all those in the various projects) and it's reached a point where
 we
  think we can start bringing on other downstream users. This means we need
  to start establishing things like a release cadence and to grow the new
  contributors we have to handle more project responsibility. Personally, I
  think that means it's time to move out from under Hadoop to drive things
 as
  our own community. Eventually, I hope the community can help draw in a
  group of folks traditionally underrepresented in ASF projects, namely QA
  and operations folks.
 
  I think test-patch by itself has enough scope to justify a project.
 Having
  a solid set of build tools that are customizable to fit the norms of
  

Re: set up jenkins test for branch-2

2015-06-15 Thread Yongjun Zhang
Thanks Sean and Allen!

I was not aware of that there is already a way to trigger branch-2 test.
Good to know.

There are multiple solutions here:

1. When posting patches, we can post two versions of patches, one for
trunk, and one for branch-2, so to trigger two pre-commit jenkins test for
both branches. This would help catching issue before committing. However,
it's going to be more workload on the testing machines, so alternatively,
we can probably defer the branch-2 testing until committing to trunk and
before branch-2 testing.

2. Only post patch for trunk, we cherry-pick to branch-2 when committing.
We can set up periodic jenkins test (such as nightly) for branch-2 to catch
problem periodically. But that's basically being ignored by us as Allen
pointed out.

The goal in my mind is to try our best to have both commits to trunk and
branch-2 to pass jenkins test, thus help on quality.

Thanks.

--Yongjun


On Mon, Jun 15, 2015 at 7:21 AM, Allen Wittenauer a...@altiscale.com wrote:


 On Jun 14, 2015, at 1:00 PM, Yongjun Zhang yzh...@cloudera.com wrote:

  From time to time we saw changes that work fine in trunk but not
 branch-2,
  and we don't catch the issue in a timely manner. The difference between
  trunk and branch-2 is sufficient to justify periodic jenkins test and
 even
  pre-commit test for branch-2.

 What is your expected outcome?

 The current nightly builds have been failing pretty consistently
 for a while now.  Adding another one isn't going to make the situation any
 better since everyone pretty much ignores them.  As I said at the HDFS
 BOF:  we desperately need a moratorium on adding new features and need to
 work on stabilizing both branch-2 and trunk.  We seem to be in a
 quantity over quality mode and it's slowly killing the project.  We
 haven't had a stable release since 2.4.1, which is approaching the 1 year
 mark.

 (Yes, I know.  Various vendors have shipped releases since 2.4.1
 as part of their distributions.  Before saying anything, perhaps you should
 spend some time with your support folks….)

 As Sean pointed out, precommit for branch-2 has been working for a
 while now if one provides enough hints to test-patch as part of the patch
 name.


[jira] [Created] (HADOOP-12090) minikdc-related unit tests fail consistently on some platforms

2015-06-15 Thread Sangjin Lee (JIRA)
Sangjin Lee created HADOOP-12090:


 Summary: minikdc-related unit tests fail consistently on some 
platforms
 Key: HADOOP-12090
 URL: https://issues.apache.org/jira/browse/HADOOP-12090
 Project: Hadoop Common
  Issue Type: Bug
  Components: kms
Affects Versions: 2.7.0
Reporter: Sangjin Lee
Assignee: Sangjin Lee


On some platforms all unit tests that use minikdc fail consistently. Those 
tests include TestKMS, TestSaslDataTransfer, TestTimelineAuthenticationFilter, 
etc.

Typical failures on the unit tests:
{noformat}
java.lang.AssertionError: 
org.apache.hadoop.security.authentication.client.AuthenticationException: 
GSSException: No valid credentials provided (Mechanism level: Cannot get a KDC 
reply)
at org.junit.Assert.fail(Assert.java:88)
at 
org.apache.hadoop.crypto.key.kms.server.TestKMS$8$4.run(TestKMS.java:1154)
at 
org.apache.hadoop.crypto.key.kms.server.TestKMS$8$4.run(TestKMS.java:1145)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1645)
at 
org.apache.hadoop.crypto.key.kms.server.TestKMS.doAs(TestKMS.java:261)
at 
org.apache.hadoop.crypto.key.kms.server.TestKMS.access$100(TestKMS.java:76)
{noformat}

The errors that cause this failure on the KDC server on the minikdc are a 
NullPointerException:
{noformat}
org.apache.mina.filter.codec.ProtocolDecoderException: 
java.lang.NullPointerException: message (Hexdump: ...)
at 
org.apache.mina.filter.codec.ProtocolCodecFilter.messageReceived(ProtocolCodecFilter.java:234)
at 
org.apache.mina.core.filterchain.DefaultIoFilterChain.callNextMessageReceived(DefaultIoFilterChain.java:434)
at 
org.apache.mina.core.filterchain.DefaultIoFilterChain.access$1200(DefaultIoFilterChain.java:48)
at 
org.apache.mina.core.filterchain.DefaultIoFilterChain$EntryImpl$1.messageReceived(DefaultIoFilterChain.java:802)
at 
org.apache.mina.core.filterchain.IoFilterAdapter.messageReceived(IoFilterAdapter.java:120)
at 
org.apache.mina.core.filterchain.DefaultIoFilterChain.callNextMessageReceived(DefaultIoFilterChain.java:434)
at 
org.apache.mina.core.filterchain.DefaultIoFilterChain.fireMessageReceived(DefaultIoFilterChain.java:426)
at 
org.apache.mina.core.polling.AbstractPollingIoProcessor.read(AbstractPollingIoProcessor.java:604)
at 
org.apache.mina.core.polling.AbstractPollingIoProcessor.process(AbstractPollingIoProcessor.java:564)
at 
org.apache.mina.core.polling.AbstractPollingIoProcessor.process(AbstractPollingIoProcessor.java:553)
at 
org.apache.mina.core.polling.AbstractPollingIoProcessor.access$400(AbstractPollingIoProcessor.java:57)
at 
org.apache.mina.core.polling.AbstractPollingIoProcessor$Processor.run(AbstractPollingIoProcessor.java:892)
at 
org.apache.mina.util.NamePreservingRunnable.run(NamePreservingRunnable.java:65)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException: message
at 
org.apache.mina.filter.codec.AbstractProtocolDecoderOutput.write(AbstractProtocolDecoderOutput.java:44)
at 
org.apache.directory.server.kerberos.protocol.codec.MinaKerberosDecoder.decode(MinaKerberosDecoder.java:65)
at 
org.apache.mina.filter.codec.ProtocolCodecFilter.messageReceived(ProtocolCodecFilter.java:224)
... 15 more
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: set up jenkins test for branch-2

2015-06-15 Thread Sean Busbey
On Mon, Jun 15, 2015 at 1:39 PM, Yongjun Zhang yzh...@cloudera.com wrote:

 Thanks Sean and Allen!

 I was not aware of that there is already a way to trigger branch-2 test.
 Good to know.

 There are multiple solutions here:

 1. When posting patches, we can post two versions of patches, one for
 trunk, and one for branch-2, so to trigger two pre-commit jenkins test for
 both branches. This would help catching issue before committing. However,
 it's going to be more workload on the testing machines, so alternatively,
 we can probably defer the branch-2 testing until committing to trunk and
 before branch-2 testing.

 2. Only post patch for trunk, we cherry-pick to branch-2 when committing.
 We can set up periodic jenkins test (such as nightly) for branch-2 to catch
 problem periodically. But that's basically being ignored by us as Allen
 pointed out.



I wouldn't worry about the load on the test machines. If you do, a third
option
is for committers to use test-patch on their local machine after making the
branch-2 version of the patch.

-- 
Sean


[jira] [Created] (HADOOP-12089) StorageException complaining no lease ID when updating FolderLastModifiedTime in WASB

2015-06-15 Thread Duo Xu (JIRA)
Duo Xu created HADOOP-12089:
---

 Summary: StorageException complaining  no lease ID when updating 
FolderLastModifiedTime in WASB
 Key: HADOOP-12089
 URL: https://issues.apache.org/jira/browse/HADOOP-12089
 Project: Hadoop Common
  Issue Type: Bug
  Components: tools
Affects Versions: 2.7.0
Reporter: Duo Xu
Assignee: Duo Xu


This is a similar issue to HADOOP-11523. HADOOP-11523 happens when HBase is 
doing distributed log splitting. This JIRA happens when HBase is deleting old 
WALs and trying to update /hbase/oldWALs folder.

The fix is the same as HADOOP-11523.

{code}
2015-06-10 08:11:40,636 WARN 
org.apache.hadoop.hbase.master.cleaner.CleanerChore: Error while deleting: 
wasb://dthbasecus...@dthbasestoragecus1.blob.core.windows.net/hbase/oldWALs/workernode10.dthbasecus1.g1.internal.cloudapp.net%2C60020%2C1433908062461.1433921692855
org.apache.hadoop.fs.azure.AzureException: 
com.microsoft.azure.storage.StorageException: There is currently a lease on the 
blob and no lease ID was specified in the request.
at 
org.apache.hadoop.fs.azure.AzureNativeFileSystemStore.updateFolderLastModifiedTime(AzureNativeFileSystemStore.java:2602)
at 
org.apache.hadoop.fs.azure.AzureNativeFileSystemStore.updateFolderLastModifiedTime(AzureNativeFileSystemStore.java:2613)
at 
org.apache.hadoop.fs.azure.NativeAzureFileSystem.delete(NativeAzureFileSystem.java:1505)
at 
org.apache.hadoop.fs.azure.NativeAzureFileSystem.delete(NativeAzureFileSystem.java:1437)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteFiles(CleanerChore.java:256)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.checkAndDeleteEntries(CleanerChore.java:157)
at 
org.apache.hadoop.hbase.master.cleaner.CleanerChore.chore(CleanerChore.java:124)
at org.apache.hadoop.hbase.Chore.run(Chore.java:80)
at java.lang.Thread.run(Thread.java:745)
Caused by: com.microsoft.azure.storage.StorageException: There is currently a 
lease on the blob and no lease ID was specified in the request.
at 
com.microsoft.azure.storage.StorageException.translateException(StorageException.java:162)
at 
com.microsoft.azure.storage.core.StorageRequest.materializeException(StorageRequest.java:307)
at 
com.microsoft.azure.storage.core.ExecutionEngine.executeWithRetry(ExecutionEngine.java:177)
at 
com.microsoft.azure.storage.blob.CloudBlob.uploadProperties(CloudBlob.java:2991)
at 
org.apache.hadoop.fs.azure.StorageInterfaceImpl$CloudBlobWrapperImpl.uploadProperties(StorageInterfaceImpl.java:372)
at 
org.apache.hadoop.fs.azure.AzureNativeFileSystemStore.updateFolderLastModifiedTime(AzureNativeFileSystemStore.java:2597)
... 8 more
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Protocol Buffers version

2015-06-15 Thread Roman Shaposhnik
On Mon, Jun 15, 2015 at 8:57 AM, Andrew Purtell apurt...@apache.org wrote:
 I can't answer the original question but can point out the protostuff (
 https://github.com/protostuff/protostuff) folks have been responsive and
 friendly in the past when we (HBase) were curious about swapping in their
 stuff. Two significant benefits of protostuff, IMHO, is ASL 2 licensing and
 everything is implemented in Java including the compiler.

Big +1 to protostuff from community, licensing and implementation perspectives.

Thanks,
Roman.


Re: Protocol Buffers version

2015-06-15 Thread Colin P. McCabe
On Mon, Jun 15, 2015 at 7:24 AM, Allen Wittenauer a...@altiscale.com wrote:

 On Jun 12, 2015, at 1:03 PM, Alan Burlison alan.burli...@oracle.com wrote:

 On 14/05/2015 18:41, Chris Nauroth wrote:

 As a reminder though, the community probably would want to see a strong
 justification for the upgrade in terms of features or performance or
 something else.  Right now, I'm not seeing a significant benefit for us
 based on my reading of their release notes.  I think it's worthwhile to
 figure this out first.  Otherwise, there is a risk that any testing work
 turns out to be a wasted effort.

 One reason at least: PB 2.5.0 has no support for Solaris SPARC. 2.6.1 does.


 That's a pretty good reason.

 Some of us had a discussion at Summit about effectively forking 
 protobuf and making it an Apache TLP.  This would give us a chance to get out 
 from under Google's blind spot, guarantee better compatibility across the 
 ecosystem, etc, etc.

 It is sounding more and more like that's really what needs to happen.

I agree that it would be nice if the protobuf project avoided making
backwards-incompatible API changes within a minor release.  But in
practice, we have had the same issues with Jackson, Guava, jets3t, and
other dependencies.  Nearly every important Hadoop dependency has made
backwards-incompatible API changes within a minor release of the
dependency... and that's one reason we are using such old versions of
everything.  I don't think PB deserves to be singled out as much as it
has been.  I think the work going on now to implement CLASSPATH
isolation in Hadoop will really be beneficial here because we will be
able to upgrade without worrying about these problems.

cheers,
Colin


Re: What is the limit to the number of properties set in the configuration object

2015-06-15 Thread Colin P. McCabe
Much like zombo.com, the only limit is yourself.

But huge Configuration objects are going to be really inefficient, so
I would look elsewhere for storing lots of data.

best,
Colin

On Fri, Jun 12, 2015 at 7:30 PM, Sitaraman Vilayannur
vrsitaramanietfli...@gmail.com wrote:
 Thanks Allen, what is the total size limit?
 Sitaraman


 On Fri, Jun 12, 2015 at 10:53 PM, Allen Wittenauer a...@altiscale.com wrote:


 On Jun 12, 2015, at 12:37 AM, Sitaraman Vilayannur 
 vrsitaramanietfli...@gmail.com wrote:

  Hi,
   What is the limit on the number of properties that can be set using
  set(String s1, String s2) on the Configuration object for hadoop?
  Is this limit configurable if so what is the maximum that can be set?

 It's a total size of the conf limit, not a number of limit.

 In general, you shouldn't pack it full of stuff as calling
 Configuration is expensive.  Use a side-input/distributed cache file for
 mass quantities of bits.





Re: 2.7.1 status

2015-06-15 Thread Vinod Kumar Vavilapalli
We are down to one blocker and a few critical tickets. I’ll try to push out an 
RC in a day or two.

Thanks
+Vinod


 On Jun 1, 2015, at 10:45 AM, Vinod Kumar Vavilapalli 
 vino...@hortonworks.com wrote:
 
 Tx for the move on that JIRA, folks.
 
 https://issues.apache.org/jira/issues/?filter=12331550 still shows 4 blockers 
 /  4 criticals. I'm going to start pushing them in/out.
 
 Thanks
 +Vinod
 
 On May 27, 2015, at 3:20 PM, Chris Nauroth cnaur...@hortonworks.com wrote:
 
 Thanks, Larry.  I have marked HADOOP-11934 as a blocker for 2.7.1.  I have
 reviewed and +1'd it.  I can commit it after we get feedback from Jenkins.
 
 --Chris Nauroth
 
 
 
 
 On 5/26/15, 12:41 PM, larry mccay lmc...@apache.org wrote:
 
 Hi Vinod -
 
 I think that https://issues.apache.org/jira/browse/HADOOP-11934 should
 also
 be added to the blocker list.
 This is a critical bug in our ability to protect the LDAP connection
 password in LdapGroupsMapper.
 
 thanks!
 
 --larry
 
 On Tue, May 26, 2015 at 3:32 PM, Vinod Kumar Vavilapalli 
 vino...@hortonworks.com wrote:
 
 Tx for reporting this, Elliot.
 
 Made it a blocker, not with a deeper understanding of the problem. Can
 you
 please chime in with your opinion and perhaps code reviews?
 
 Thanks
 +Vinod
 
 On May 26, 2015, at 10:48 AM, Elliott Clark ecl...@apache.org wrote:
 
 HADOOP-12001 should probably be added to the blocker list since it's a
 regression that can keep ldap from working.
 
 
 
 
 
 



Re: CMake and CMAKE_LD_FLAGS

2015-06-15 Thread Colin P. McCabe
Hi Alan,

I think you are right that CMAKE_LD_FLAGS has never done anything, and
setting it was always a mistake.

The uses of CMAKE_LD_FLAGS in JNIFlags.cmake are harmless, since the
-m32 option is not needed for the linker.  However, it's more
concerning that hadoop-mapreduce-client-nativetask seems to be using
this macro to set -no-undefined -version-info 0:1:0 and some other
stuff.

Can you create a JIRA to change these uses over to the correct CMake
directive?  If you create a JIRA that makes only that change, it
should be pretty quick to review and commit.

best,
Colin

On Thu, Jun 4, 2015 at 7:41 AM, Alan Burlison alan.burli...@oracle.com wrote:
 On 03/06/2015 21:23, Alan Burlison wrote:

 hadoop-common-project/hadoop-common/src/JNIFlags.cmake sets the
 CMAKE_LD_FLAGS variable to the flags required for the linker. However,
 despite the name, CMAKE_LD_FLAGS is *not* a valid standard CMake
 variable


 Confirmed, I've put nonsense flags in CMAKE_LD_FLAGS that should produce an
 invalid linker command-line and they don't appear in the generated makefiles
 and the build works fine. If I put the same nonsense flags in the proper
 variables then the build blows up as I'd expect.

 --
 Alan Burlison
 --


Re: Maven always detects changes - Is this a Docker 'feature'?

2015-06-15 Thread Colin P. McCabe
Hi Darrell,

Sorry, I'm not familiar with this feature of Maven.  Perhaps try
asking on the Apache Maven mailing list?

best,
Colin

On Fri, May 22, 2015 at 8:34 AM, Darrell Taylor
darrell.tay...@gmail.com wrote:
 Hi,

 Is it normal behaviour for maven to detect changes when I run tests with no
 changes?

 e.g.
 $ mvn test -Dtest=TestDFSShell -nsu -o
 ...
 [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @
 hadoop-hdfs ---
 [INFO] Changes detected - recompiling the module!
 [INFO] Compiling 576 source files to
 /home/darrell/hadoop/hadoop-hdfs-project/hadoop-hdfs/target/classes
 ...

 Then I run the same command again without touching anything else and it
 compiles everything again.  It's getting rather tedious.

 I am running this from inside the docker container.

 Any help appreciated.

 Thanks
 Darrell.


Re: 2.7.1 status

2015-06-15 Thread Tsuyoshi Ozawa
Hi Vinod,

Thank you for working to release 2.7!
YARN-3798 looks a critical issue for branch-2.7. I'll mark it as a
blocker of 2.7.1 if possible.

Thanks,
- Tsuyoshi

On Tue, Jun 16, 2015 at 6:58 AM, Vinod Kumar Vavilapalli
vino...@hortonworks.com wrote:
 We are down to one blocker and a few critical tickets. I’ll try to push out 
 an RC in a day or two.

 Thanks
 +Vinod


 On Jun 1, 2015, at 10:45 AM, Vinod Kumar Vavilapalli 
 vino...@hortonworks.com wrote:

 Tx for the move on that JIRA, folks.

 https://issues.apache.org/jira/issues/?filter=12331550 still shows 4 
 blockers /  4 criticals. I'm going to start pushing them in/out.

 Thanks
 +Vinod

 On May 27, 2015, at 3:20 PM, Chris Nauroth cnaur...@hortonworks.com wrote:

 Thanks, Larry.  I have marked HADOOP-11934 as a blocker for 2.7.1.  I have
 reviewed and +1'd it.  I can commit it after we get feedback from Jenkins.

 --Chris Nauroth




 On 5/26/15, 12:41 PM, larry mccay lmc...@apache.org wrote:

 Hi Vinod -

 I think that https://issues.apache.org/jira/browse/HADOOP-11934 should
 also
 be added to the blocker list.
 This is a critical bug in our ability to protect the LDAP connection
 password in LdapGroupsMapper.

 thanks!

 --larry

 On Tue, May 26, 2015 at 3:32 PM, Vinod Kumar Vavilapalli 
 vino...@hortonworks.com wrote:

 Tx for reporting this, Elliot.

 Made it a blocker, not with a deeper understanding of the problem. Can
 you
 please chime in with your opinion and perhaps code reviews?

 Thanks
 +Vinod

 On May 26, 2015, at 10:48 AM, Elliott Clark ecl...@apache.org wrote:

 HADOOP-12001 should probably be added to the blocker list since it's a
 regression that can keep ldap from working.









Re: [DISCUSS] project for pre-commit patch testing (was Re: upstream jenkins build broken?)

2015-06-15 Thread Sean Busbey
Oof. I had meant to push on this again but life got in the way and now the
June board meeting is upon us. Sorry everyone. In the event that this ends
up contentious, hopefully one of the copied communities can give us a
branch to work in.

I know everyone is busy, so here's the short version of this email: I'd
like to move some of the code currently in Hadoop (test-patch) into a new
TLP focused on QA tooling. I'm not sure what the best format for priming
this conversation is. ORC filled in the incubator project proposal
template, but I'm not sure how much that confused the issue. So to start,
I'll just write what I'm hoping we can accomplish in general terms here.

All software development projects that are community based (that is,
accepting outside contributions) face a common QA problem for vetting
in-coming contributions. Hadoop is fortunate enough to be sufficiently
popular that the weight of the problem drove tool development (i.e.
test-patch). That tool is generalizable enough that a bunch of other TLPs
have adopted their own forks. Unfortunately, in most projects this kind of
QA work is an enabler rather than a primary concern, so often the tooling
is worked on ad-hoc and little shared improvements happen across
projects. Since
the tooling itself is never a primary concern, any made is rarely reused
outside of ASF projects.

Over the last couple months a few of us have been working on generalizing
the tooling present in the Hadoop code base (because it was the most mature
out of all those in the various projects) and it's reached a point where we
think we can start bringing on other downstream users. This means we need
to start establishing things like a release cadence and to grow the new
contributors we have to handle more project responsibility. Personally, I
think that means it's time to move out from under Hadoop to drive things as
our own community. Eventually, I hope the community can help draw in a
group of folks traditionally underrepresented in ASF projects, namely QA
and operations folks.

I think test-patch by itself has enough scope to justify a project. Having
a solid set of build tools that are customizable to fit the norms of
different software communities is a bunch of work. Making it work well in
both the context of automated test systems like Jenkins and for individual
developers is even more work. We could easily also take over maintenance of
things like shelldocs, since test-patch is the primary consumer of that
currently but it's generally useful tooling.

In addition to test-patch, I think the proposed project has some future
growth potential. Given some adoption of test-patch to prove utility, the
project could build on the ties it makes to start building tools to help
projects do their own longer-run testing. Note that I'm talking about the
tools to build QA processes and not a particular set of tested components.
Specifically, I think the ChaosMonkey work that's in HBase should be
generalizable as a fault injection framework (either based on that code or
something like it). Doing this for arbitrary software is obviously very
difficult, and a part of easing that will be to make (and then favor)
tooling to allow projects to have operational glue that looks the same.
Namely, the shell work that's been done in hadoop-functions.sh would be a
great foundational layer that could bring good daemon handling practices to
a whole slew of software projects. In the event that these frameworks and
tools get adopted by parts of the Hadoop ecosystem, that could make the job
of i.e. Bigtop substantially easier.

I've reached out to a few folks who have been involved in the current
test-patch work or expressed interest in helping out on getting it used in
other projects. Right now, the proposed PMC would be (alphabetical by last
name):

* Andrew Bayer (ASF member, incubator pmc, bigtop pmc, flume pmc, jclouds
pmc, sqoop pmc, all around Jenkins expert)
* Sean Busbey (ASF member, accumulo pmc, hbase pmc)
* Nick Dimiduk (hbase pmc, phoenix pmc)
* Chris Nauroth (ASF member, incubator pmc, hadoop pmc)
* Andrew Purtell  (ASF member, incubator pmc, bigtop pmc, hbase pmc,
phoenix pmc)
* Allen Wittenauer (hadoop committer)

That PMC gives us several members and a bunch of folks familiar with the
ASF. Combined with the code already existing in Apache spaces, I think that
gives us sufficient justification for a direct board proposal.

The planned project name is Apache Yetus. It's an archaic genus of sea
snail and most of our project will be focused on shell scripts.

N.b.: this does not mean that the Hadoop community would _have_ to rely on
the new TLP, but I hope that once we have a release that can be evaluated
there'd be enough benefit to strongly encourage it.

This has mostly been focused on scope and community issues, and I'd love to
talk through any feedback on that. Additionally, are there any other points
folks want to make sure are covered before we have a resolution?

On Sat, Jun 6, 2015