Success: YARN-1452 PreCommit Build #3452

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1452
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3452/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3042 lines...]

/bin/kill -9 6326 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636524/YARN-1452.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+0 tests included{color}.  The patch appears to be a 
documentation patch that doesn't require tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3452//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3452//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
177dd5657406014d35ec98d720a1d1172ebbe506 logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1452
Recording test results
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

Build failed in Jenkins: Hadoop-Yarn-trunk #520

2014-03-25 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Yarn-trunk/520/changes

Changes:

[jing9] HDFS-5840. Follow-up to HDFS-5138 to improve error handling during 
partial upgrade failures. Contributed by Aaron T. Myers, Suresh Srinivas, and 
Jing Zhao.

[suresh] HDFS-6125. Cleanup unnecessary cast in HDFS code base. Contributed by 
Suresh Srinivas.

[vinodkv] YARN-1850. Introduced the ability to optionally disable sending out 
timeline-events in the TimelineClient. Contributed by Zhijie Shen.

[szetszwo] HADOOP-10425. LocalFileSystem.getContentSummary should not count crc 
files.

[vinodkv] MAPREDUCE-5795. Fixed MRAppMaster to record the correct job-state 
after it recovers from a commit during a previous attempt. Contributed by Xuan 
Gong.

[arp] HDFS-6124. Add final modifier to class members. (Contributed by Suresh 
Srinivas)

[kasha] HADOOP-10423. Clarify compatibility policy document for combination of 
new client and old server. (Chris Nauroth via kasha)

[cnauroth] HADOOP-10422. Remove redundant logging of RPC retry attempts. 
Contributed by Chris Nauroth.

[cnauroth] HDFS-5846. Shuffle phase is slow in Windows - 
FadviseFileRegion::transferTo does not read disks efficiently. Contributed by 
Nikola Vujic.

[jing9] Move HDFS-5138 to 2.4.0 section in CHANGES.txt

[jing9] HDFS-6135. In HDFS upgrade with HA setup, JournalNode cannot handle 
layout version bump when rolling back. Contributed by Jing Zhao.

[brandonli] HDFS-6050. NFS does not handle exceptions correctly in a few 
places. Contributed by Brandon Li

[jianhe] YARN-1852. Fixed RMAppAttempt to not resend 
AttemptFailed/AttemptKilled events to already recovered Failed/Killed RMApps. 
Contributed by Rohith Sharmaks

[cnauroth] MAPREDUCE-5791. Shuffle phase is slow in Windows - 
FadviseFileRegion::transferTo does not read disks efficiently. Contributed by 
Nikola Vujic.

[szetszwo] HADOOP-10015. UserGroupInformation prints out excessive warnings.  
Contributed by Nicolas Liochon

[zjshen] YARN-1838. Enhanced timeline service getEntities API to get entities 
from a given entity ID or insertion timestamp. Contributed by Billie Rinaldi.

[jeagles] YARN-1670. aggregated log writer can write more log data then it says 
is the log length (Mit Desai via jeagles)

[kihwal] HDFS-3087. Decomissioning on NN restart can complete without blocks 
being replicated. Contributed by Rushabh S Shah.

--
[...truncated 18060 lines...]
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:198)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025)
at 
java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:424)
at 
org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2065)
IPC Server handler 4 on 59853 daemon prio=5 tid=238 timed_waiting
java.lang.Thread.State: TIMED_WAITING
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:198)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025)
at 
java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:424)
at 
org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2065)
IPC Server handler 7 on 45253 daemon prio=5 tid=174 timed_waiting
java.lang.Thread.State: TIMED_WAITING
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:198)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025)
at 
java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:424)
at 
org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2065)
IPC Server handler 8 on 59853 daemon prio=5 tid=242 timed_waiting
java.lang.Thread.State: TIMED_WAITING
at sun.misc.Unsafe.park(Native Method)
at 
java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:198)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:2025)
at 
java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:424)
at 
org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:109)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2065)
IPC Server Responder daemon prio=5 tid=233 runnable
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:210)
at 

[jira] [Created] (YARN-1871) We should eliminate writing *PBImpl code in YARN

2014-03-25 Thread Wangda Tan (JIRA)
Wangda Tan created YARN-1871:


 Summary: We should eliminate writing *PBImpl code in YARN
 Key: YARN-1871
 URL: https://issues.apache.org/jira/browse/YARN-1871
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: api
Affects Versions: 2.4.0
Reporter: Wangda Tan
Assignee: Wangda Tan



Currently, We need write PBImpl classes one by one. After running find . -name 
*PBImpl*.java | xargs wc -l under hadoop source code directory, we can see, 
there're more than 25,000 LOC. I think we should improve this, which will be 
very helpful for YARN developers to make changes for YARN protocols.


There're only some limited patterns in current *PBImpl,
* Simple types, like string, int32, float.
* List? types
* Map? types
* Enum types
Code generation should be enough to generate such PBImpl classes.

Some other requirements are,
* Leave other related code alone, like service implemention (e.g. 
ContainerManagerImpl).
* (If possible) Forward compatibility, developpers can write their own PBImpl 
or genereate them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1872) TestDistributedShell occasionally fails in trunk

2014-03-25 Thread Ted Yu (JIRA)
Ted Yu created YARN-1872:


 Summary: TestDistributedShell occasionally fails in trunk
 Key: YARN-1872
 URL: https://issues.apache.org/jira/browse/YARN-1872
 Project: Hadoop YARN
  Issue Type: Test
Reporter: Ted Yu


From https://builds.apache.org/job/Hadoop-Yarn-trunk/520/console :

TestDistributedShell#testDSShellWithCustomLogPropertyFile failed and 
TestDistributedShell#testDSShell timed out.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Re: About Map 100% reduce %0 issue

2014-03-25 Thread Hitesh Shah
You are missing the following in your yarn site:

property
  nameyarn.nodemanager.aux-services.mapreduce_shuffle.class/name
  valueorg.apache.hadoop.mapred.ShuffleHandler/value
/property

You will need to restart the nodemanager for this property to take effect. 

-- Hitesh

On Mar 24, 2014, at 9:34 PM, Vincent,Wei wrote:

 All
 
 I am a new comer for Hadoop, I have run
 the hadoop-mapreduce-examples-2.2.0.jar wordcount, but the result is that
 it always pending at map 100% and reduce %0.
 
 14/03/25 20:19:20 INFO client.RMProxy: Connecting to ResourceManager at
 master/159.99.249.63:8032
 14/03/25 20:19:20 INFO input.FileInputFormat: Total input paths to process
 : 1
 14/03/25 20:19:20 INFO mapreduce.JobSubmitter: number of splits:1
 14/03/25 20:19:20 INFO Configuration.deprecation: user.name is deprecated.
 Instead, use mapreduce.job.user.name
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.jar is deprecated.
 Instead, use mapreduce.job.jar
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.output.value.class
 is deprecated. Instead, use mapreduce.job.output.value.class
 14/03/25 20:19:20 INFO Configuration.deprecation: mapreduce.combine.class
 is deprecated. Instead, use mapreduce.job.combine.class
 14/03/25 20:19:20 INFO Configuration.deprecation: mapreduce.map.class is
 deprecated. Instead, use mapreduce.job.map.class
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.job.name is
 deprecated. Instead, use mapreduce.job.name
 14/03/25 20:19:20 INFO Configuration.deprecation: mapreduce.reduce.class is
 deprecated. Instead, use mapreduce.job.reduce.class
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.input.dir is
 deprecated. Instead, use mapreduce.input.fileinputformat.inputdir
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.output.dir is
 deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.map.tasks is
 deprecated. Instead, use mapreduce.job.maps
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.output.key.class
 is deprecated. Instead, use mapreduce.job.output.key.class
 14/03/25 20:19:20 INFO Configuration.deprecation: mapred.working.dir is
 deprecated. Instead, use mapreduce.job.working.dir
 14/03/25 20:19:20 INFO mapreduce.JobSubmitter: Submitting tokens for job:
 job_1395747600383_0002
 14/03/25 20:19:20 INFO impl.YarnClientImpl: Submitted application
 application_1395747600383_0002 to ResourceManager at master/
 159.99.249.63:8032
 14/03/25 20:19:20 INFO mapreduce.Job: The url to track the job:
 http://master:8088/proxy/application_1395747600383_0002/
 14/03/25 20:19:20 INFO mapreduce.Job: Running job: job_1395747600383_0002
 14/03/25 20:19:24 INFO mapreduce.Job: Job job_1395747600383_0002 running in
 uber mode : false
 14/03/25 20:19:24 INFO mapreduce.Job:  map 0% reduce 0%
 14/03/25 20:19:28 INFO mapreduce.Job:  map 100% reduce 0%
 14/03/25 20:19:31 INFO mapreduce.Job: Task Id :
 attempt_1395747600383_0002_r_00_0, Status : FAILED
 Error: org.apache.hadoop.mapreduce.task.reduce.Shuffle$ShuffleError: error
 in shuffle in fetcher#5
at
 org.apache.hadoop.mapreduce.task.reduce.Shuffle.run(Shuffle.java:121)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:380)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
 Caused by: java.io.IOException: Exceeded MAX_FAILED_UNIQUE_FETCHES;
 bailing-out.
at
 org.apache.hadoop.mapreduce.task.reduce.ShuffleSchedulerImpl.checkReducerHealth(ShuffleSchedulerImpl.java:323)
at
 org.apache.hadoop.mapreduce.task.reduce.ShuffleSchedulerImpl.copyFailed(ShuffleSchedulerImpl.java:245)
at
 org.apache.hadoop.mapreduce.task.reduce.Fetcher.copyFromHost(Fetcher.java:347)
at
 org.apache.hadoop.mapreduce.task.reduce.Fetcher.run(Fetcher.java:165)
 
 someone says that this is caused by hosts configure .I have checked my
 /etc/hosts  on all Mater  slaves:
 127.0.0.1   localhost.localdomain localhost
 159.99.249.63   master
 159.99.249.203  slave1
 159.99.249.99   slave2
 159.99.249.88   slave3
 
 Would you please help me to fix the issue, many thanks .
 
 my yarn-site.xml
 
 ?xml version=1.0?
 ?xml-stylesheet type=text/xsl href=configuration.xsl?
 
 
 configuration
 
 property
 descriptionThe hostname of the RM./description
 nameyarn.resourcemanager.hostname/name
 valuemaster/value
 /property
 
 property
 nameyarn.nodemanager.aux-services/name
 valuemapreduce_shuffle/value
 /property
 
 property
 descriptionThe address of the container manager in the NM./description
 nameyarn.nodemanager.address/name
 value${yarn.nodemanager.hostname}:8041/value
 /property
 
 

[jira] [Created] (YARN-1873) TestDistributedShell#testDSShell fails

2014-03-25 Thread Mit Desai (JIRA)
Mit Desai created YARN-1873:
---

 Summary: TestDistributedShell#testDSShell fails
 Key: YARN-1873
 URL: https://issues.apache.org/jira/browse/YARN-1873
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Mit Desai
Assignee: Mit Desai


testDSShell fails when the tests are run in random order. I see a cleanup issue 
here.

{noformat}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 72.222 sec  
FAILURE! - in 
org.apache.hadoop.yarn.applications.distributedshell.TestDistributedShell
testOrder(org.apache.hadoop.yarn.applications.distributedshell.TestDistributedShell)
  Time elapsed: 44.127 sec   FAILURE!
java.lang.AssertionError: expected:1 but was:6
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at org.junit.Assert.assertEquals(Assert.java:456)
at 
org.apache.hadoop.yarn.applications.distributedshell.TestDistributedShell.testDSShell(TestDistributedShell.java:204)
at 
org.apache.hadoop.yarn.applications.distributedshell.TestDistributedShell.testOrder(TestDistributedShell.java:134)


Results :

Failed tests: 
  TestDistributedShell.testOrder:134-testDSShell:204 expected:1 but was:6
{noformat}

The Line numbers will be little deviated because I was trying to reproduce the 
error by running the tests in specific order. But the Line that causes the 
assert fail is {{Assert.assertEquals(1, 
entitiesAttempts.getEntities().size());}}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Failed: YARN-1867 PreCommit Build #3454

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1867
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3454/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 123 lines...]
no change for http://svn.apache.org/repos/asf/hadoop/nightly since the previous 
build
No emails were triggered.
[PreCommit-YARN-Build] $ /bin/bash /tmp/hudson2505182573111833833.sh
32768
Running in Jenkins mode


==
==
Testing patch for YARN-1867.
==
==


At revision 1581491.
YARN-1867 patch is being downloaded at Tue Mar 25 20:16:10 UTC 2014 from
http://issues.apache.org/jira/secure/attachment/12636769/YARN-1867-20140325.txt
cp: cannot stat `/home/jenkins/buildSupport/lib/*': No such file or directory
The patch does not appear to apply with p0 to p2
PATCH APPLICATION FAILED




{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12636769/YARN-1867-20140325.txt
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3454//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
a13a58a4953899dce318ffa54beb82e02f4858d2 logged out


==
==
Finished build.
==
==


Build step 'Execute shell' marked build as failure
Archiving artifacts
[description-setter] Could not determine description.
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
No tests ran.

[jira] [Created] (YARN-1874) Cleanup: Move RMActiveServices out of ResourceManager into its own file

2014-03-25 Thread Karthik Kambatla (JIRA)
Karthik Kambatla created YARN-1874:
--

 Summary: Cleanup: Move RMActiveServices out of ResourceManager 
into its own file
 Key: YARN-1874
 URL: https://issues.apache.org/jira/browse/YARN-1874
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Reporter: Karthik Kambatla


As [~vinodkv] noticed on YARN-1867, ResourceManager is hard to maintain. We 
should move RMActiveServices out to make it more manageable. 





--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1875) TestRMHA#testStartAndTransitions is failing

2014-03-25 Thread Mit Desai (JIRA)
Mit Desai created YARN-1875:
---

 Summary: TestRMHA#testStartAndTransitions is failing
 Key: YARN-1875
 URL: https://issues.apache.org/jira/browse/YARN-1875
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Mit Desai
 Attachments: Log.rtf

{noformat}
testStartAndTransitions(org.apache.hadoop.yarn.server.resourcemanager.TestRMHA) 
 Time elapsed: 5.883 sec   FAILURE!
java.lang.AssertionError: Incorrect value for metric availableMB 
expected:2048 but was:4096
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.failNotEquals(Assert.java:647)
at org.junit.Assert.assertEquals(Assert.java:128)
at org.junit.Assert.assertEquals(Assert.java:472)
at 
org.apache.hadoop.yarn.server.resourcemanager.TestRMHA.assertMetric(TestRMHA.java:396)
at 
org.apache.hadoop.yarn.server.resourcemanager.TestRMHA.verifyClusterMetrics(TestRMHA.java:387)
at 
org.apache.hadoop.yarn.server.resourcemanager.TestRMHA.testStartAndTransitions(TestRMHA.java:160)


Results :

Failed tests: 
  
TestRMHA.testStartAndTransitions:160-verifyClusterMetrics:387-assertMetric:396
 Incorrect value for metric availableMB expected:2048 but was:4096
{noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Success: YARN-1521 PreCommit Build #3453

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1521
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3453/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3480 lines...]

/bin/kill -9 32335 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636766/YARN-1521.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3453//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3453//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
ffb0999be98ef3777ae37f6bbe56d4063467d725 logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1521
Recording test results
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

Success: YARN-1873 PreCommit Build #3455

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1873
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3455/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 2930 lines...]
/bin/kill -9 24014 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636768/YARN-1873.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3455//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3455//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
a1eb818b527f31d694d66bacc54cc12dfc264e2a logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1873
Recording test results
Publish JUnit test result report is waiting for a checkpoint on 
PreCommit-YARN-Build #3453
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

Success: YARN-1867 PreCommit Build #3458

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1867
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3458/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3036 lines...]

/bin/kill -9 31907 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12636787/YARN-1867-20140325-trunk.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3458//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3458//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
6b3bdb58d8504b6e6d941ab60bd64f75b7417fe5 logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1867
Recording test results
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

Success: YARN-1521 PreCommit Build #3459

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1521
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3459/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3367 lines...]

/bin/kill -9 10434 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636836/YARN-1521.3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 6 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3459//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3459//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
a69787a04bd77fd8b25bafd250153ccb2c6f2527 logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1521
Recording test results
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

Success: YARN-1452 PreCommit Build #3460

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1452
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3460/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 2989 lines...]

/bin/kill -9 12563 
kill: No such process
NOP




{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636842/YARN-1452.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+0 tests included{color}.  The patch appears to be a 
documentation patch that doesn't require tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3460//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3460//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
45cb610b14d20bf66ef6dfacb93ad3a60fd1cdb3 logged out


==
==
Finished build.
==
==


Archiving artifacts
Description set: YARN-1452
Recording test results
Email was triggered for: Success
Sending email for trigger: Success



###
## FAILED TESTS (if any) 
##
All tests passed

[jira] [Created] (YARN-1876) Document the REST APIs of timeline and generic history services

2014-03-25 Thread Zhijie Shen (JIRA)
Zhijie Shen created YARN-1876:
-

 Summary: Document the REST APIs of timeline and generic history 
services
 Key: YARN-1876
 URL: https://issues.apache.org/jira/browse/YARN-1876
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Zhijie Shen
Assignee: Zhijie Shen






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1877) ZK store: Add yarn.resourcemanager.zk-state-store.root-node.auth for root node auth

2014-03-25 Thread Karthik Kambatla (JIRA)
Karthik Kambatla created YARN-1877:
--

 Summary: ZK store: Add 
yarn.resourcemanager.zk-state-store.root-node.auth for root node auth
 Key: YARN-1877
 URL: https://issues.apache.org/jira/browse/YARN-1877
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.3.0
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
Priority: Critical






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1878) Yarn standby RM taking long to transition to active

2014-03-25 Thread Arpit Gupta (JIRA)
Arpit Gupta created YARN-1878:
-

 Summary: Yarn standby RM taking long to transition to active
 Key: YARN-1878
 URL: https://issues.apache.org/jira/browse/YARN-1878
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Arpit Gupta
Assignee: Xuan Gong


In our HA tests we are noticing that some times it can take upto 10s for the 
standby RM to transition to active.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


Failed: YARN-1878 PreCommit Build #3461

2014-03-25 Thread Apache Jenkins Server
Jira: https://issues.apache.org/jira/browse/YARN-1878
Build: https://builds.apache.org/job/PreCommit-YARN-Build/3461/

###
## LAST 60 LINES OF THE CONSOLE 
###
[...truncated 3076 lines...]
NOP




{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12636856/YARN-1878.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3461//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3461//console

This message is automatically generated.


==
==
Adding comment to Jira.
==
==


Comment added.
7f5d27cea31c7828e476b12f68f3ea3a2a486be5 logged out


==
==
Finished build.
==
==


Build step 'Execute shell' marked build as failure
Archiving artifacts
[description-setter] Could not determine description.
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



###
## FAILED TESTS (if any) 
##
All tests passed

RE: Capacity scheduler puts all containers on one box

2014-03-25 Thread Brahma Reddy Battula
As I knew,,I never seen RM assigning containers entirely one NM until that NM 
is fullI might be wrong...

But I used to configure following three properties,which will restirict the 
number of containers for one NM ..


Hope following configurations will help you to solve your disk saturation and 
following formula is obtained based on some tests.

# of containers = min (2*CORES, 1.8*DISKS, (Total available RAM) / 
MIN_CONTAINER_SIZE)


property
nameyarn.nodemanager.resource.memory-mb/name
value40960/value
  /property

 property
namemapreduce.map.memory.mb/name
value4096/value
  /property
  property
namemapreduce.reduce.memory.mb/name
value8192/value
  /property



Thanks  Regards
Brahma Reddy Battula



From: Sandy Ryza [sandy.r...@cloudera.com]
Sent: Saturday, March 22, 2014 12:18 PM
To: yarn-dev@hadoop.apache.org
Subject: Re: Capacity scheduler puts all containers on one box

For a work-around, if you turn on DRF / multi-resource scheduling, you
could use vcore capacities to limit the number of containers per node?

On Fri, Mar 21, 2014 at 11:35 PM, Chris Riccomini
criccom...@linkedin.comwrote:

 Hey Guys,

 @Vinod: We aren't overriding the default, so we must be using -1 as the
 setting.

 @Sandy: We aren't specifying any racks/hosts when sending the resource
 requests. +1 regarding introducing a similar limit in capacity scheduler.

 Any recommended work-arounds in the mean time? Our utilization of the grid
 is very low because we're having to force high memory requests for the
 containers in order to guarantee a maximum number of containers on a
 single node (e.g. Set container memory MB set to 17GB to disallow more
 than 2 containers from being assigned to any one 48GB node).

 Cheers,
 Chris

 On 3/21/14 11:30 PM, Sandy Ryza sandy.r...@cloudera.com wrote:

 yarn.scheduler.capacity.node-locality-delay will help if the app is
 requesting containers at particular locations, but won't help spread
 things
 out evenly otherwise.
 
 The Fair Scheduler attempts an even spread.  By default, it only schedules
 a single container each time it considers a node.  Decoupling scheduling
 from node heartbeats (YARN-1010) makes it so that a high node heartbeat
 interval doesn't result in this being slow.  Now that the Capacity
 Scheduler has similar capabilities (YARN-1512), it might make sense to
 introduce a similar limit?
 
 -Sandy
 
 
 On Fri, Mar 21, 2014 at 4:42 PM, Vinod Kumar Vavilapalli
 vino...@apache.org
  wrote:
 
  What's the value for yarn.scheduler.capacity.node-locality-delay? It is
 -1
  by default in 2.2.
 
  We fixed the default to be a reasonable 40 (nodes in a rack) in 2.3.0
 that
  should spread containers a bit.
 
  Thanks,
  +Vinod
 
  On Mar 21, 2014, at 12:48 PM, Chris Riccomini criccom...@linkedin.com
  wrote:
 
   Hey Guys,
  
   We're running YARN 2.2 with the capacity scheduler. Each NM is running
  with 40G of memory capacity. When we request a series containers with
 2G of
  memory from a single AM, we see the RM assigning them entirely to one NM
  until that NM is full, and then moving on to the next, and so on.
  Essentially, we have a grid with 20 nodes, and two are completely full,
 and
  the rest are completely empty. This is problematic because our
 containers
  use disk heavily, and are completely saturating the disks on the two
 nodes,
  which slows all of the containers down on these NMs.
  
1.  Is this expected behavior of the capacity scheduler? What about
 the
  fifo scheduler?
2.  Is the recommended work around just to increase memory allocation
  per-container as a proxy for the disk capacity that's required? Given
 that
  there's no disk-level isolation, and no disk-level resource, I don't see
  another way around this.
  
   Cheers,
   Chris
 
 
  --
  CONFIDENTIALITY NOTICE
  NOTICE: This message is intended for the use of the individual or
 entity to
  which it is addressed and may contain information that is confidential,
  privileged and exempt from disclosure under applicable law. If the
 reader
  of this message is not the intended recipient, you are hereby notified
 that
  any printing, copying, dissemination, distribution, disclosure or
  forwarding of this communication is strictly prohibited. If you have
  received this communication in error, please contact the sender
 immediately
  and delete it from your system. Thank You.
 



Re: Capacity scheduler puts all containers on one box

2014-03-25 Thread Gordon Wang
Hi Chris,

I agree with Sandy, if you want to distribute container evenly, fair
scheduler is a much better choice.

If you insist to use capacity scheduler, as Vinod said, if setting
yarn.scheduler.capacity.node-locality-delay to the # node in each rack,
this will spread the containers to different nodes.

One more thing you have to check is the data distribution. If your data is
distributed evenly in the cluster, setting
yarn.scheduler.capacity.node-locality-delay
is useful. But if your data is skewed, the containers will be assigned to
part of the nodes due to data locality awareness.


On Sat, Mar 22, 2014 at 2:35 PM, Chris Riccomini criccom...@linkedin.comwrote:

 Hey Guys,

 @Vinod: We aren't overriding the default, so we must be using -1 as the
 setting.

 @Sandy: We aren't specifying any racks/hosts when sending the resource
 requests. +1 regarding introducing a similar limit in capacity scheduler.

 Any recommended work-arounds in the mean time? Our utilization of the grid
 is very low because we're having to force high memory requests for the
 containers in order to guarantee a maximum number of containers on a
 single node (e.g. Set container memory MB set to 17GB to disallow more
 than 2 containers from being assigned to any one 48GB node).

 Cheers,
 Chris

 On 3/21/14 11:30 PM, Sandy Ryza sandy.r...@cloudera.com wrote:

 yarn.scheduler.capacity.node-locality-delay will help if the app is
 requesting containers at particular locations, but won't help spread
 things
 out evenly otherwise.
 
 The Fair Scheduler attempts an even spread.  By default, it only schedules
 a single container each time it considers a node.  Decoupling scheduling
 from node heartbeats (YARN-1010) makes it so that a high node heartbeat
 interval doesn't result in this being slow.  Now that the Capacity
 Scheduler has similar capabilities (YARN-1512), it might make sense to
 introduce a similar limit?
 
 -Sandy
 
 
 On Fri, Mar 21, 2014 at 4:42 PM, Vinod Kumar Vavilapalli
 vino...@apache.org
  wrote:
 
  What's the value for yarn.scheduler.capacity.node-locality-delay? It is
 -1
  by default in 2.2.
 
  We fixed the default to be a reasonable 40 (nodes in a rack) in 2.3.0
 that
  should spread containers a bit.
 
  Thanks,
  +Vinod
 
  On Mar 21, 2014, at 12:48 PM, Chris Riccomini criccom...@linkedin.com
  wrote:
 
   Hey Guys,
  
   We're running YARN 2.2 with the capacity scheduler. Each NM is running
  with 40G of memory capacity. When we request a series containers with
 2G of
  memory from a single AM, we see the RM assigning them entirely to one NM
  until that NM is full, and then moving on to the next, and so on.
  Essentially, we have a grid with 20 nodes, and two are completely full,
 and
  the rest are completely empty. This is problematic because our
 containers
  use disk heavily, and are completely saturating the disks on the two
 nodes,
  which slows all of the containers down on these NMs.
  
1.  Is this expected behavior of the capacity scheduler? What about
 the
  fifo scheduler?
2.  Is the recommended work around just to increase memory allocation
  per-container as a proxy for the disk capacity that's required? Given
 that
  there's no disk-level isolation, and no disk-level resource, I don't see
  another way around this.
  
   Cheers,
   Chris
 
 
  --
  CONFIDENTIALITY NOTICE
  NOTICE: This message is intended for the use of the individual or
 entity to
  which it is addressed and may contain information that is confidential,
  privileged and exempt from disclosure under applicable law. If the
 reader
  of this message is not the intended recipient, you are hereby notified
 that
  any printing, copying, dissemination, distribution, disclosure or
  forwarding of this communication is strictly prohibited. If you have
  received this communication in error, please contact the sender
 immediately
  and delete it from your system. Thank You.
 




-- 
Regards
Gordon Wang


[jira] [Created] (YARN-1879) Mark Idempotent/AtMostOnce annotations to ApplicationMasterProtocol and add tests for it

2014-03-25 Thread Jian He (JIRA)
Jian He created YARN-1879:
-

 Summary: Mark Idempotent/AtMostOnce annotations to 
ApplicationMasterProtocol and add tests for it
 Key: YARN-1879
 URL: https://issues.apache.org/jira/browse/YARN-1879
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Jian He






--
This message was sent by Atlassian JIRA
(v6.2#6252)