[jira] [Created] (MAPREDUCE-6343) JobConf.parseMaximumHeapSizeMB() fails to parse value greater than 2GB expressed in bytes

2015-04-27 Thread Hao Xia (JIRA)
Hao Xia created MAPREDUCE-6343:
--

 Summary: JobConf.parseMaximumHeapSizeMB() fails to parse value 
greater than 2GB expressed in bytes
 Key: MAPREDUCE-6343
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6343
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Hao Xia


It currently tries to parse the value as an integer, which blows up whenever 
the value is greater than 2GB and expressed in bytes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Planning Hadoop 2.6.1 release

2015-04-27 Thread Vinod Kumar Vavilapalli
There were several requests on the user lists [1] for a 2.6.1 release. I
got many offline comments too.

Planning to do a 2.6.1 release in a few weeks time. We already have a bunch
of tickets committed to 2.7.1. I created a filter [2] to tracking pending
tickets.

We need to collectively come up with a list of critical issues. We can use
the JIRA Target Version field for the same. I see some but not a whole lot
of new work for this release, most of it is likely going to be pulling in
critical patches from 2.7.1/2.8 etc.

Thoughts?

Thanks
+Vinod

[1] Will Hadoop 2.6.1 be released soon?
http://markmail.org/thread/zlsr6prejyogdyvh
[2] 2.6.1 pending tickets
https://issues.apache.org/jira/issues/?filter=12331711


Re: [DISCUSS] Release numbering for stable 2.8 and beyond

2015-04-27 Thread Sandy Ryza
My understanding was that the main reason that we labeled 2.0 alpha and 2.1
beta is that we wanted flexibility to make breaking API changes.

Is that the case now with 2.7?  I.e. do we have APIs labeled as Public /
Stable that we want freedom to change in 2.8?  If not, I definitely don't
think the release merits a special tag.  We should just make sure its bugs
are well documented.

-Sandy

On Mon, Apr 27, 2015 at 10:45 AM, Andrew Wang 
wrote:

> I think Karthik correctly identified the two reasons we might want to
> denote a release as "unstable":
>
> a) Compatibility. Whether we have freedom to break compatibility before the
> final release, i.e. what "alpha" typically means.
>
> b) Quality. As Konst says, a release can be allowed to break compatibility
> ("alpha") but itself still be a high quality release.
>
> We could try and separate out these two concerns when it comes to
> versioning, but I think in reality users only care about prod vs. non-prod,
> which is why many other projects (Linux, HBase, etc) just have two release
> lines: "stable" (compatible/good-quality) vs. "unstable"
> (unknown-compatible/unknown-quality).
>
> To this end, I don't mind what we choose, as long as the difference between
> stable and unstable is denoted by the version number. I don't like the idea
> of tagging a release as good after the fact (1). The other projects we've
> used as examples make strong statements about their stable release lines,
> and we should too. Our downstreams and end users will appreciate clarity
> from the version number.
>
> Best,
> Andrew
>
> On Sun, Apr 26, 2015 at 9:51 AM, Hitesh Shah  wrote:
>
> > There are a couple of different approaches we could take.
> >
> > How about publishing/releasing bits such as “2.8.0-RC0”. Downstream
> > projects can depend on these and use them normally similar to the
> approach
> > that they would have taken with release 2.8.0 or 2.8.0-alpha. After some
> > baking ( or more RC suffixed releases), at some point, a proper 2.8.0
> > release could be made? The RC tag clearly indicates to downstream layers
> > that things will be in flux slightly. Trying to distinguish
> > incompatibilities between 2.8.0 and 2.8.1/2.8.2 ( just because 2.8.0 was
> > tagged alpha in documentation ) are likely to be a nightmare to deal with
> > especially for new features introduced in the 2.8 release ( which might
> get
> > changed slightly based on usage feedback ).
> >
> > Furthermore, considering the release history of hadoop, the likelihood
> > that 2.9.0 will show up before 2.8.2 seems to be very high.
> >
> > With respect to the proposed choice (1), I thought that HBase applied a
> > different approach. The odd-number releases were always unstable and the
> > even number releases were the stable ones. This “could" make things a bit
> > more clear for downstream users without needing to resort to modified
> > versioned numbers ( with alpha/beta tags ) or requiring users to go look
> up
> > the website to find out which particular version of 2.8.x was the first
> > stable release.
> >
> > thanks
> > — Hitesh
> >
> >
> >
> >
> >
> > On Apr 22, 2015, at 2:17 PM, Vinod Kumar Vavilapalli <
> > vino...@hortonworks.com> wrote:
> >
> > > Forking the thread.
> > >
> > > In the previous 2.7.1 thread [1], there were enough yays to my proposal
> > to wait for a bug-fix release or two before calling a 2.x release stable.
> > There were some concerns about the naming.
> > >
> > > We have two options, taking 2.8 as an example
> > > (1) Release 2.8.0, call it as an alpha in documentation and release
> > notes, wait for a 2.8.1/2.8.2 reasonably stable enough to be called as
> the
> > first stable release of 2.8.
> > > (2) Release 2.8.0-alpha, 2.8.0-beta etc before culminating in a 2.8.0
> > stable release.
> > >
> > > (1) is what I preferred first up. This is what HBase used to do, and
> far
> > beyond, in the linux kernel releases. It helps in scenarios where we are
> > forced to downgrade a release, say due to major issues. We can simply
> > announce it as not stable retroactively, change the pointers on our
> website
> > and move on.
> > >
> > > Thoughts?
> > >
> > > Thanks,
> > > +Vinod
> > >
> > > [1] http://markmail.org/thread/ogzk4phj6wsdpssu
> > >
> > > On Apr 21, 2015, at 4:59 PM, Vinod Kumar Vavilapalli <
> > vino...@hortonworks.com> wrote:
> > >
> > >>
> > >> Sure, I agree it's better to have clear guidelines and scheme. Let me
> > fork this thread about that.
> > >>
> > >> Re 2.7.0, I just forgot about the naming initially though I was clear
> > in the discussion/voting. I so had to end up calling it alpha outside of
> > the release artifact naming.
> > >>
> > >> Thanks
> > >> +Vinod
> > >>
> > >> On Apr 21, 2015, at 4:26 PM, Andrew Wang 
> > wrote:
> > >>
> > >>> I would also like to support Karthik's proposal on the release thread
> > about
> > >>> version numbering. 2.7.0 being "alpha" is pretty confusing since all
> > of the
> > >>> other 2.x releases since GA have been stable. I 

[jira] [Created] (MAPREDUCE-6342) Make POM project names consistent

2015-04-27 Thread Rohith (JIRA)
Rohith created MAPREDUCE-6342:
-

 Summary: Make POM project names consistent
 Key: MAPREDUCE-6342
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6342
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: build
Reporter: Rohith
Assignee: Rohith
Priority: Minor


This is track MR changes for POM changes  by name



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [DISCUSS] Release numbering for stable 2.8 and beyond

2015-04-27 Thread Andrew Wang
I think Karthik correctly identified the two reasons we might want to
denote a release as "unstable":

a) Compatibility. Whether we have freedom to break compatibility before the
final release, i.e. what "alpha" typically means.

b) Quality. As Konst says, a release can be allowed to break compatibility
("alpha") but itself still be a high quality release.

We could try and separate out these two concerns when it comes to
versioning, but I think in reality users only care about prod vs. non-prod,
which is why many other projects (Linux, HBase, etc) just have two release
lines: "stable" (compatible/good-quality) vs. "unstable"
(unknown-compatible/unknown-quality).

To this end, I don't mind what we choose, as long as the difference between
stable and unstable is denoted by the version number. I don't like the idea
of tagging a release as good after the fact (1). The other projects we've
used as examples make strong statements about their stable release lines,
and we should too. Our downstreams and end users will appreciate clarity
from the version number.

Best,
Andrew

On Sun, Apr 26, 2015 at 9:51 AM, Hitesh Shah  wrote:

> There are a couple of different approaches we could take.
>
> How about publishing/releasing bits such as “2.8.0-RC0”. Downstream
> projects can depend on these and use them normally similar to the approach
> that they would have taken with release 2.8.0 or 2.8.0-alpha. After some
> baking ( or more RC suffixed releases), at some point, a proper 2.8.0
> release could be made? The RC tag clearly indicates to downstream layers
> that things will be in flux slightly. Trying to distinguish
> incompatibilities between 2.8.0 and 2.8.1/2.8.2 ( just because 2.8.0 was
> tagged alpha in documentation ) are likely to be a nightmare to deal with
> especially for new features introduced in the 2.8 release ( which might get
> changed slightly based on usage feedback ).
>
> Furthermore, considering the release history of hadoop, the likelihood
> that 2.9.0 will show up before 2.8.2 seems to be very high.
>
> With respect to the proposed choice (1), I thought that HBase applied a
> different approach. The odd-number releases were always unstable and the
> even number releases were the stable ones. This “could" make things a bit
> more clear for downstream users without needing to resort to modified
> versioned numbers ( with alpha/beta tags ) or requiring users to go look up
> the website to find out which particular version of 2.8.x was the first
> stable release.
>
> thanks
> — Hitesh
>
>
>
>
>
> On Apr 22, 2015, at 2:17 PM, Vinod Kumar Vavilapalli <
> vino...@hortonworks.com> wrote:
>
> > Forking the thread.
> >
> > In the previous 2.7.1 thread [1], there were enough yays to my proposal
> to wait for a bug-fix release or two before calling a 2.x release stable.
> There were some concerns about the naming.
> >
> > We have two options, taking 2.8 as an example
> > (1) Release 2.8.0, call it as an alpha in documentation and release
> notes, wait for a 2.8.1/2.8.2 reasonably stable enough to be called as the
> first stable release of 2.8.
> > (2) Release 2.8.0-alpha, 2.8.0-beta etc before culminating in a 2.8.0
> stable release.
> >
> > (1) is what I preferred first up. This is what HBase used to do, and far
> beyond, in the linux kernel releases. It helps in scenarios where we are
> forced to downgrade a release, say due to major issues. We can simply
> announce it as not stable retroactively, change the pointers on our website
> and move on.
> >
> > Thoughts?
> >
> > Thanks,
> > +Vinod
> >
> > [1] http://markmail.org/thread/ogzk4phj6wsdpssu
> >
> > On Apr 21, 2015, at 4:59 PM, Vinod Kumar Vavilapalli <
> vino...@hortonworks.com> wrote:
> >
> >>
> >> Sure, I agree it's better to have clear guidelines and scheme. Let me
> fork this thread about that.
> >>
> >> Re 2.7.0, I just forgot about the naming initially though I was clear
> in the discussion/voting. I so had to end up calling it alpha outside of
> the release artifact naming.
> >>
> >> Thanks
> >> +Vinod
> >>
> >> On Apr 21, 2015, at 4:26 PM, Andrew Wang 
> wrote:
> >>
> >>> I would also like to support Karthik's proposal on the release thread
> about
> >>> version numbering. 2.7.0 being "alpha" is pretty confusing since all
> of the
> >>> other 2.x releases since GA have been stable. I think users would
> prefer a
> >>> version like "2.8.0-alpha1" instead, which is very clear and similar to
> >>> what we did for 2.0 and 2.1. Then we release 2.8.0 when we're actually
> >>> stable.
> >>>
> >>> I don't know if it's retroactively possible to do this for 2.7.0, but
> it's
> >>> something to consider for the next 2.7 alpha or beta or whatever.
> >>>
> >
>
>


Hadoop-Mapreduce-trunk - Build # 2126 - Failure

2015-04-27 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2126/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 31993 lines...]
Results :

Tests in error: 
  
TestClusterMRNotification>NotificationTestCase.setUp:145->HadoopTestCase.setUp:157
 » YarnRuntime

Tests run: 520, Failures: 0, Errors: 1, Skipped: 11

[INFO] 
[INFO] Reactor Summary:
[INFO] 
[INFO] hadoop-mapreduce-client ... SUCCESS [  2.714 s]
[INFO] hadoop-mapreduce-client-core .. SUCCESS [01:51 min]
[INFO] hadoop-mapreduce-client-common  SUCCESS [ 30.935 s]
[INFO] hadoop-mapreduce-client-shuffle ... SUCCESS [  4.679 s]
[INFO] hadoop-mapreduce-client-app ... SUCCESS [12:54 min]
[INFO] hadoop-mapreduce-client-hs  SUCCESS [05:59 min]
[INFO] hadoop-mapreduce-client-jobclient . FAILURE [  01:45 h]
[INFO] hadoop-mapreduce-client-hs-plugins  SKIPPED
[INFO] hadoop-mapreduce-client-nativetask  SKIPPED
[INFO] Apache Hadoop MapReduce Examples .. SKIPPED
[INFO] hadoop-mapreduce .. SKIPPED
[INFO] 
[INFO] BUILD FAILURE
[INFO] 
[INFO] Total time: 02:07 h
[INFO] Finished at: 2015-04-27T15:34:40+00:00
[INFO] Final Memory: 34M/779M
[INFO] 
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-mapreduce-client-jobclient: There are test failures.
[ERROR] 
[ERROR] Please refer to 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire-reports
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn  -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Sending artifact delta relative to Hadoop-Mapreduce-trunk #2125
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 20329379 bytes
Compression is 0.0%
Took 6.8 sec
Recording test results
Updating HADOOP-11857
Updating YARN-3464
Updating MAPREDUCE-6057
Updating HADOOP-11865
Updating HADOOP-11357
Updating MAPREDUCE-6252
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
1 tests failed.
REGRESSION:  org.apache.hadoop.mapred.TestClusterMRNotification.testMR

Error Message:
java.lang.NoClassDefFoundError: 
org/apache/hadoop/yarn/server/MiniYARNCluster$NodeManagerWrapper$1

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: 
java.lang.NoClassDefFoundError: 
org/apache/hadoop/yarn/server/MiniYARNCluster$NodeManagerWrapper$1
at 
org.apache.hadoop.yarn.server.MiniYARNCluster$NodeManagerWrapper.serviceStart(MiniYARNCluster.java:576)
at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
at 
org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
at 
org.apache.hadoop.mapred.MiniMRClientClusterFactory.create(MiniMRClientClusterFactory.java:80)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:187)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:175)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:167)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:159)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:152)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:145)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:138)
at org.apache.hadoop.mapred.MiniMRCluster.(MiniMRCluster.java:133)
at 
org.apache.hadoop.mapred.HadoopTestCase.setUp(HadoopTestCase.java:157)
at 
org.apache.hadoop.mapred.NotificationTestC

Hadoop-Mapreduce-trunk-Java8 - Build # 177 - Still Failing

2015-04-27 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/177/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 26974 lines...]
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.808 sec - in 
org.apache.hadoop.mapred.TestReporter

Results :

Failed tests: 
  TestCombineFileInputFormat.testSplitPlacementForCompressedFiles:918 
expected:<2> but was:<1>
  TestCombineFileInputFormat.testSplitPlacement:372 expected:<2> but was:<1>

Tests run: 509, Failures: 2, Errors: 0, Skipped: 11

[INFO] 
[INFO] Reactor Summary:
[INFO] 
[INFO] hadoop-mapreduce-client ... SUCCESS [  2.091 s]
[INFO] hadoop-mapreduce-client-core .. SUCCESS [01:38 min]
[INFO] hadoop-mapreduce-client-common  SUCCESS [ 24.876 s]
[INFO] hadoop-mapreduce-client-shuffle ... SUCCESS [  4.009 s]
[INFO] hadoop-mapreduce-client-app ... SUCCESS [13:23 min]
[INFO] hadoop-mapreduce-client-hs  SUCCESS [07:14 min]
[INFO] hadoop-mapreduce-client-jobclient . FAILURE [  02:00 h]
[INFO] hadoop-mapreduce-client-hs-plugins  SKIPPED
[INFO] hadoop-mapreduce-client-nativetask  SKIPPED
[INFO] Apache Hadoop MapReduce Examples .. SKIPPED
[INFO] hadoop-mapreduce .. SKIPPED
[INFO] 
[INFO] BUILD FAILURE
[INFO] 
[INFO] Total time: 02:22 h
[INFO] Finished at: 2015-04-27T15:30:30+00:00
[INFO] Final Memory: 33M/192M
[INFO] 
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-mapreduce-client-jobclient: There was a timeout or other error 
in the fork -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn  -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Sending artifact delta relative to Hadoop-Mapreduce-trunk-Java8 #28
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 20341150 bytes
Compression is 0.0%
Took 8.4 sec
Recording test results
Updating HADOOP-11857
Updating YARN-3464
Updating MAPREDUCE-6057
Updating HADOOP-11865
Updating HADOOP-11357
Updating MAPREDUCE-6252
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
2 tests failed.
FAILED:  
org.apache.hadoop.mapreduce.lib.input.TestCombineFileInputFormat.testSplitPlacementForCompressedFiles

Error Message:
expected:<2> but was:<1>

Stack Trace:
java.lang.AssertionError: expected:<2> but was:<1>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.mapreduce.lib.input.TestCombineFileInputFormat.testSplitPlacementForCompressedFiles(TestCombineFileInputFormat.java:918)


FAILED:  
org.apache.hadoop.mapreduce.lib.input.TestCombineFileInputFormat.testSplitPlacement

Error Message:
expected:<2> but was:<1>

Stack Trace:
java.lang.AssertionError: expected:<2> but was:<1>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.mapreduce.lib.input.TestCombineFileInputFormat.testSplitPlacement(TestCombineFileInputFormat.java:372)




[jira] [Created] (MAPREDUCE-6340) Remove .xml and documentation references to dfs.webhdfs.enabled

2015-04-27 Thread Ray Chiang (JIRA)
Ray Chiang created MAPREDUCE-6340:
-

 Summary: Remove .xml and documentation references to 
dfs.webhdfs.enabled
 Key: MAPREDUCE-6340
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6340
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Ray Chiang
Assignee: Ray Chiang
Priority: Minor
 Fix For: 3.0.0


HDFS-7985 removed any references in the code to the dfs.webhdfs.enabled 
property.  The property should also be removed from hdfs-default.xml as well as 
other places.

As mentioned in HDFS-7985, this is an incompatible change with branch-2, so 
this fix should be limited to trunk.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)