[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755412#comment-13755412
 ] 

Hadoop QA commented on YARN-1065:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600917/YARN-1065.5.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 4 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1814//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1814//console

This message is automatically generated.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch, YARN-1065.5.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755382#comment-13755382
 ] 

Xuan Gong commented on YARN-1065:
-

Thanks for the comments.
Expose the ContainerManagerImpl to containerLauncher and containerLaunch, that 
we can use ContainerManagerImpl.auxiliaryServices.getMetaData() to get all the 
service data

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch, YARN-1065.5.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1065:


Attachment: YARN-1065.5.patch

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch, YARN-1065.5.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1090) Job does not get into Pending State

2013-08-30 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755368#comment-13755368
 ] 

Sandy Ryza commented on YARN-1090:
--

Ah, ok, makes total sense

> Job does not get into Pending State
> ---
>
> Key: YARN-1090
> URL: https://issues.apache.org/jira/browse/YARN-1090
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: yeshavora
>Assignee: Jian He
> Attachments: YARN-1090.patch
>
>
> When there is no resource available to run a job, next job should go in 
> pending state. RM UI should show next job as pending app and the counter for 
> the pending app should be incremented.
> But Currently. Next job stays in ACCEPTED state and No AM has been assigned 
> to this job.Though Pending App count is not incremented. 
> Running 'job status ' shows job state=PREP. 
> $ mapred job -status job_1377122233385_0002
> 13/08/21 21:59:23 INFO client.RMProxy: Connecting to ResourceManager at 
> host1/ip1
> Job: job_1377122233385_0002
> Job File: /ABC/.staging/job_1377122233385_0002/job.xml
> Job Tracking URL : http://host1:port1/application_1377122233385_0002/
> Uber job : false
> Number of maps: 0
> Number of reduces: 0
> map() completion: 0.0
> reduce() completion: 0.0
> Job state: PREP
> retired: false
> reason for failure:

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755334#comment-13755334
 ] 

Bikas Saha commented on YARN-1065:
--

We need the service data provided by the Auxillary Service and not the service 
data sent by the AM in the container launch context. The AM provides 
application part of service data. In response, the aux service provides the 
service part of the service data. This service part == 
ContainerManagerImpl.auxiliaryServices.getMetaData(). That is what we need to 
set into the env of the container via sanitizeEnv. 

It would be great if we use matching names like getServiceDataFromEnv() and 
setServiceDataIntoEnv().

Would be good to prepend something NM specific to the service name before 
setting in the env. This just helps identify that this was set by the NM. 
Something like NM_AUX_SERVICE_. So for service abcd, we will add 
NM_AUX_SERVICE_abcd into the env.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1116) Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755329#comment-13755329
 ] 

Hadoop QA commented on YARN-1116:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600898/YARN-1116.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1813//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1813//console

This message is automatically generated.

> Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts
> 
>
> Key: YARN-1116
> URL: https://issues.apache.org/jira/browse/YARN-1116
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
> Attachments: YARN-1116.patch
>
>
> The AMRMTokens are now only saved in RMStateStore and not populated back to 
> AMRMTokenSecretManager after RM restarts. This is more needed now since 
> AMRMToken also becomes used in non-secure env.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-771) AMRMClient support for resource blacklisting

2013-08-30 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755313#comment-13755313
 ] 

Junping Du commented on YARN-771:
-

Thanks [~bikassaha] for review!

> AMRMClient  support for resource blacklisting
> -
>
> Key: YARN-771
> URL: https://issues.apache.org/jira/browse/YARN-771
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Bikas Saha
>Assignee: Junping Du
> Fix For: 2.1.1-beta
>
> Attachments: YARN-771-v1.0.patch, YARN-771-v2.patch, 
> YARN-771-v3.patch, YARN-771-v4.patch
>
>
> After YARN-750 AMRMClient should support blacklisting via the new YARN API's

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1116) Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts

2013-08-30 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755305#comment-13755305
 ] 

Jian He commented on YARN-1116:
---

Tested on single node cluster, now the AMRMToken can be recognized by RM after 
it restarts.

> Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts
> 
>
> Key: YARN-1116
> URL: https://issues.apache.org/jira/browse/YARN-1116
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
> Attachments: YARN-1116.patch
>
>
> The AMRMTokens are now only saved in RMStateStore and not populated back to 
> AMRMTokenSecretManager after RM restarts. This is more needed now since 
> AMRMToken also becomes used in non-secure env.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755303#comment-13755303
 ] 

Hadoop QA commented on YARN-1065:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600896/YARN-1065.4.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1812//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1812//console

This message is automatically generated.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1116) Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts

2013-08-30 Thread Jian He (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1116?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jian He updated YARN-1116:
--

Attachment: YARN-1116.patch

Upload a patch that populates AMRMToken's password back to 
AMRMTokenSecretManager after RM restarts.

> Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts
> 
>
> Key: YARN-1116
> URL: https://issues.apache.org/jira/browse/YARN-1116
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
> Attachments: YARN-1116.patch
>
>
> The AMRMTokens are now only saved in RMStateStore and not populated back to 
> AMRMTokenSecretManager after RM restarts. This is more needed now since 
> AMRMToken also becomes used in non-secure env.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1065:


Attachment: YARN-1065.4.patch

Fix -1 on findBug
Modify the test case to test if environment variable is forwarded using 
sanitizeEnv

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch, 
> YARN-1065.4.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1116) Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts

2013-08-30 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755278#comment-13755278
 ] 

Jian He commented on YARN-1116:
---

Looks like simply populating the password inside AMRMToken back to 
AMRMTokenSecretManager after RM restarts is enough.

> Populate AMRMTokens back to AMRMTokenSecretManager after RM restarts
> 
>
> Key: YARN-1116
> URL: https://issues.apache.org/jira/browse/YARN-1116
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
>
> The AMRMTokens are now only saved in RMStateStore and not populated back to 
> AMRMTokenSecretManager after RM restarts. This is more needed now since 
> AMRMToken also becomes used in non-secure env.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (YARN-1133) AMRMClient should be able to blacklist nodes, and specify them in AllocateRequest

2013-08-30 Thread Zhijie Shen (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1133?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijie Shen resolved YARN-1133.
---

Resolution: Duplicate

> AMRMClient should be able to blacklist nodes, and specify them in 
> AllocateRequest
> -
>
> Key: YARN-1133
> URL: https://issues.apache.org/jira/browse/YARN-1133
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Zhijie Shen
>Assignee: Zhijie Shen
>
> After YARN-750, YARN scheduler is able to blacklist nodes when scheduling. 
> AMRMClient should enable this feature to the clients.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1117) Improve help message for $ yarn applications and $yarn node

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755265#comment-13755265
 ] 

Hudson commented on YARN-1117:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4355 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4355/])
YARN-1117. Improved help messages for "yarn application" and "yarn node" 
commands. Contributed by Xuan Gong. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1519117)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/ApplicationCLI.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/NodeCLI.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestYarnClient.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ConverterUtils.java


> Improve help message for $ yarn applications and $yarn node
> ---
>
> Key: YARN-1117
> URL: https://issues.apache.org/jira/browse/YARN-1117
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1117.1.patch, YARN-1117.2.patch, YARN-1117.3.patch
>
>
> There is standardization of help message in YARN-1080. It is nice to have 
> similar changes for $ yarn appications and yarn node

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1117) Improve help message for $ yarn applications and $yarn node

2013-08-30 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1117?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755253#comment-13755253
 ] 

Vinod Kumar Vavilapalli commented on YARN-1117:
---

Looks good, +1. Checking this in.

> Improve help message for $ yarn applications and $yarn node
> ---
>
> Key: YARN-1117
> URL: https://issues.apache.org/jira/browse/YARN-1117
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1117.1.patch, YARN-1117.2.patch, YARN-1117.3.patch
>
>
> There is standardization of help message in YARN-1080. It is nice to have 
> similar changes for $ yarn appications and yarn node

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1133) AMRMClient should be able to blacklist nodes, and specify them in AllocateRequest

2013-08-30 Thread Zhijie Shen (JIRA)
Zhijie Shen created YARN-1133:
-

 Summary: AMRMClient should be able to blacklist nodes, and specify 
them in AllocateRequest
 Key: YARN-1133
 URL: https://issues.apache.org/jira/browse/YARN-1133
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Zhijie Shen
Assignee: Zhijie Shen


After YARN-750, YARN scheduler is able to blacklist nodes when scheduling. 
AMRMClient should enable this feature to the clients.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-696) Enable multiple states to to be specified in Resource Manager apps REST call

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755031#comment-13755031
 ] 

Hadoop QA commented on YARN-696:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600827/YARN-696.diff
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1807//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1807//console

This message is automatically generated.

> Enable multiple states to to be specified in Resource Manager apps REST call
> 
>
> Key: YARN-696
> URL: https://issues.apache.org/jira/browse/YARN-696
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Affects Versions: 2.0.4-alpha
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: YARN-696.diff, YARN-696.diff
>
>
> Within the YARN Resource Manager REST API the GET call which returns all 
> Applications can be filtered by a single State query parameter (http:// http address:port>/ws/v1/cluster/apps). 
> There are 8 possible states (New, Submitted, Accepted, Running, Finishing, 
> Finished, Failed, Killed), if no state parameter is specified all states are 
> returned, however if a sub-set of states is required then multiple REST calls 
> are required (max. of 7).
> The proposal is to be able to specify multiple states in a single REST call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-888) clean up POM dependencies

2013-08-30 Thread Timothy St. Clair (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755236#comment-13755236
 ] 

Timothy St. Clair commented on YARN-888:


Our current list of JIRA's can be found here: 
https://fedoraproject.org/wiki/Changes/Hadoop#Upstream_patch_tracking

> clean up POM dependencies
> -
>
> Key: YARN-888
> URL: https://issues.apache.org/jira/browse/YARN-888
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.1.0-beta
>Reporter: Alejandro Abdelnur
>Assignee: Roman Shaposhnik
>
> Intermediate 'pom' modules define dependencies inherited by leaf modules.
> This is causing issues in intellij IDE.
> We should normalize the leaf modules like in common, hdfs and tools where all 
> dependencies are defined in each leaf module and the intermediate 'pom' 
> module do not define any dependency.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-771) AMRMClient support for resource blacklisting

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755235#comment-13755235
 ] 

Hudson commented on YARN-771:
-

SUCCESS: Integrated in Hadoop-trunk-Commit #4354 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4354/])
YARN-771. AMRMClient support for resource blacklisting (Junping Du via bikas) 
(bikas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1519107)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/AMRMClient.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/impl/AMRMClientImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/ResourceBlacklistRequestPBImpl.java


> AMRMClient  support for resource blacklisting
> -
>
> Key: YARN-771
> URL: https://issues.apache.org/jira/browse/YARN-771
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Bikas Saha
>Assignee: Junping Du
> Attachments: YARN-771-v1.0.patch, YARN-771-v2.patch, 
> YARN-771-v3.patch, YARN-771-v4.patch
>
>
> After YARN-750 AMRMClient should support blacklisting via the new YARN API's

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755228#comment-13755228
 ] 

Hadoop QA commented on YARN-1065:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600867/YARN-1065.3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 1 new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1811//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-YARN-Build/1811//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-yarn-server-nodemanager.html
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1811//console

This message is automatically generated.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1090) Job does not get into Pending State

2013-08-30 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755227#comment-13755227
 ] 

Jian He commented on YARN-1090:
---

bq. My understanding was that an application is "Pending" if no AM has yet been 
allocated to it
Yes, that's correct. What I meant here is about the UI of the Application 
Queues usage inside which:
'Num Pending Applications' doesn't mean that no AM has been allocated it, and 
propose to rename it to 'Num non-schedulable applications'.

> Job does not get into Pending State
> ---
>
> Key: YARN-1090
> URL: https://issues.apache.org/jira/browse/YARN-1090
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: yeshavora
>Assignee: Jian He
> Attachments: YARN-1090.patch
>
>
> When there is no resource available to run a job, next job should go in 
> pending state. RM UI should show next job as pending app and the counter for 
> the pending app should be incremented.
> But Currently. Next job stays in ACCEPTED state and No AM has been assigned 
> to this job.Though Pending App count is not incremented. 
> Running 'job status ' shows job state=PREP. 
> $ mapred job -status job_1377122233385_0002
> 13/08/21 21:59:23 INFO client.RMProxy: Connecting to ResourceManager at 
> host1/ip1
> Job: job_1377122233385_0002
> Job File: /ABC/.staging/job_1377122233385_0002/job.xml
> Job Tracking URL : http://host1:port1/application_1377122233385_0002/
> Uber job : false
> Number of maps: 0
> Number of reduces: 0
> map() completion: 0.0
> reduce() completion: 0.0
> Job state: PREP
> retired: false
> reason for failure:

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1090) Job does not get into Pending State

2013-08-30 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755219#comment-13755219
 ] 

Sandy Ryza commented on YARN-1090:
--

My understanding was that an application is "Pending" if no AM has yet been 
allocated to it.  Can you go a little more into what the issue is with this 
definition?

Also, whatever we decide, is there somewhere where we can document exactly what 
these mean?

> Job does not get into Pending State
> ---
>
> Key: YARN-1090
> URL: https://issues.apache.org/jira/browse/YARN-1090
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: yeshavora
>Assignee: Jian He
> Attachments: YARN-1090.patch
>
>
> When there is no resource available to run a job, next job should go in 
> pending state. RM UI should show next job as pending app and the counter for 
> the pending app should be incremented.
> But Currently. Next job stays in ACCEPTED state and No AM has been assigned 
> to this job.Though Pending App count is not incremented. 
> Running 'job status ' shows job state=PREP. 
> $ mapred job -status job_1377122233385_0002
> 13/08/21 21:59:23 INFO client.RMProxy: Connecting to ResourceManager at 
> host1/ip1
> Job: job_1377122233385_0002
> Job File: /ABC/.staging/job_1377122233385_0002/job.xml
> Job Tracking URL : http://host1:port1/application_1377122233385_0002/
> Uber job : false
> Number of maps: 0
> Number of reduces: 0
> map() completion: 0.0
> reduce() completion: 0.0
> Job state: PREP
> retired: false
> reason for failure:

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1132) QueueMetrics.java has wrong comments

2013-08-30 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755208#comment-13755208
 ] 

Jian He commented on YARN-1132:
---

Hi [~ajisakaa], thanks for reporting this.
This will be fixed as part of YARN-1090, thx!

> QueueMetrics.java has wrong comments
> 
>
> Key: YARN-1132
> URL: https://issues.apache.org/jira/browse/YARN-1132
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: documentation
>Affects Versions: 2.1.0-beta
>Reporter: Akira AJISAKA
>Priority: Minor
>  Labels: newbie
>
> I found o.a.h.yarn.server.resourcemanager.scheduler.QueueMetrics.java has 
> wrong comments
> {code}
>   @Metric("# of reserved memory in MB") MutableGaugeInt reservedMB;
>   @Metric("# of active users") MutableGaugeInt activeApplications;
> {code}
> they should be fixed as follows:
> {code}
>   @Metric("Reserved memory in MB") MutableGaugeInt reservedMB;
>   @Metric("# of active applications") MutableGaugeInt activeApplications;
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1132) QueueMetrics.java has wrong comments

2013-08-30 Thread Akira AJISAKA (JIRA)
Akira AJISAKA created YARN-1132:
---

 Summary: QueueMetrics.java has wrong comments
 Key: YARN-1132
 URL: https://issues.apache.org/jira/browse/YARN-1132
 Project: Hadoop YARN
  Issue Type: Bug
  Components: documentation
Affects Versions: 2.1.0-beta
Reporter: Akira AJISAKA
Priority: Minor


I found o.a.h.yarn.server.resourcemanager.scheduler.QueueMetrics.java has wrong 
comments

{code}
  @Metric("# of reserved memory in MB") MutableGaugeInt reservedMB;
  @Metric("# of active users") MutableGaugeInt activeApplications;
{code}

they should be fixed as follows:

{code}
  @Metric("Reserved memory in MB") MutableGaugeInt reservedMB;
  @Metric("# of active applications") MutableGaugeInt activeApplications;
{code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1131) $ yarn logs should return a message log aggregation is during progress if YARN application is running

2013-08-30 Thread Tassapol Athiapinya (JIRA)
Tassapol Athiapinya created YARN-1131:
-

 Summary: $ yarn logs should return a message log aggregation is 
during progress if YARN application is running
 Key: YARN-1131
 URL: https://issues.apache.org/jira/browse/YARN-1131
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: client
Reporter: Tassapol Athiapinya
Priority: Minor
 Fix For: 2.1.1-beta


In the case when log aggregation is enabled, if a user submits MapReduce job 
and runs $ yarn logs -applicationId  while the YARN application is 
running, the command will return no message and return user back to shell. It 
is nice to tell the user that log aggregation is in progress.

{code}
-bash-4.1$ /usr/bin/yarn logs -applicationId application_1377900193583_0002
-bash-4.1$
{code}

At the same time, if invalid application ID is given, YARN CLI should say that 
the application ID is incorrect rather than throwing NoSuchElementException.
{code}
$ /usr/bin/yarn logs -applicationId application_0
Exception in thread "main" java.util.NoSuchElementException
at com.google.common.base.AbstractIterator.next(AbstractIterator.java:75)
at 
org.apache.hadoop.yarn.util.ConverterUtils.toApplicationId(ConverterUtils.java:124)
at 
org.apache.hadoop.yarn.util.ConverterUtils.toApplicationId(ConverterUtils.java:119)
at org.apache.hadoop.yarn.logaggregation.LogDumper.run(LogDumper.java:110)
at org.apache.hadoop.yarn.logaggregation.LogDumper.main(LogDumper.java:255)

{code}


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1130) Improve the log flushing for tasks when mapred.userlog.limit.kb is set

2013-08-30 Thread Paul Han (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1130?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Paul Han updated YARN-1130:
---

Description: 
When userlog limit is set with something like this:
{code}

mapred.userlog.limit.kb
2048
The maximum size of user-logs of each task in KB. 0 disables the 
cap.


{code}
the log entry will be truncated randomly for the jobs.

The log size is left between 1.2MB to 1.6MB.

Since the log is already limited, avoid the log truncation is crucial for user.

The other issue with the current 
impl(org.apache.hadoop.yarn.ContainerLogAppender) is that log entries will not 
flush to file until the container shutdown and logmanager close all appenders. 
If user likes to see the log during task execution, it doesn't support it.

Will propose a patch to add a flush mechanism and also flush the log when task 
is done.  


  was:
When userlog limit is set with something like this:
{code}

mapred.userlog.limit.kb
2048
The maximum size of user-logs of each task in KB. 0 disables the 
cap.


{code}
the log entry will be truncated randomly for the jobs.

The log size is left between 1.2MB to 1.6MB.

Since the log is already limited, avoid the log truncation is crucial for user.

The other issue with the current 
impl(org.apache.hadoop.yarn.ContainerLogAppender) will not write to log until 
the container shutdown and logmanager close all appenders. If user likes to see 
the log during task execution, it doesn't support it.

Will propose a patch to add a flush mechanism and also flush the log when task 
is done.  



> Improve the log flushing for tasks when mapred.userlog.limit.kb is set
> --
>
> Key: YARN-1130
> URL: https://issues.apache.org/jira/browse/YARN-1130
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 2.0.5-alpha
>Reporter: Paul Han
>
> When userlog limit is set with something like this:
> {code}
> 
> mapred.userlog.limit.kb
> 2048
> The maximum size of user-logs of each task in KB. 0 disables the 
> cap.
> 
> 
> {code}
> the log entry will be truncated randomly for the jobs.
> The log size is left between 1.2MB to 1.6MB.
> Since the log is already limited, avoid the log truncation is crucial for 
> user.
> The other issue with the current 
> impl(org.apache.hadoop.yarn.ContainerLogAppender) is that log entries will 
> not flush to file until the container shutdown and logmanager close all 
> appenders. If user likes to see the log during task execution, it doesn't 
> support it.
> Will propose a patch to add a flush mechanism and also flush the log when 
> task is done.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1130) Improve the log flushing for tasks when mapred.userlog.limit.kb is set

2013-08-30 Thread Paul Han (JIRA)
Paul Han created YARN-1130:
--

 Summary: Improve the log flushing for tasks when 
mapred.userlog.limit.kb is set
 Key: YARN-1130
 URL: https://issues.apache.org/jira/browse/YARN-1130
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Affects Versions: 2.0.5-alpha
Reporter: Paul Han


When userlog limit is set with something like this:
{code}

mapred.userlog.limit.kb
2048
The maximum size of user-logs of each task in KB. 0 disables the 
cap.


{code}
the log entry will be truncated randomly for the jobs.

The log size is left between 1.2MB to 1.6MB.

Since the log is already limited, avoid the log truncation is crucial for user.

The other issue with the current 
impl(org.apache.hadoop.yarn.ContainerLogAppender) will not write to log until 
the container shutdown and logmanager close all appenders. If user likes to see 
the log during task execution, it doesn't support it.

Will propose a patch to add a flush mechanism and also flush the log when task 
is done.  


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1065:


Attachment: YARN-1065.3.patch

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch, YARN-1065.3.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1098) Separate out RM services into "Always On" and "Active"

2013-08-30 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1098:
---

Attachment: yarn-1098-1.patch

Uploading a patch along the lines of my previous commit.

The only change is that AdminService is also in the Active services. We should 
move it to "Always On" once we make it HA-aware; currently, some of the 
functions don't apply to Standby RM.

Also, the patch depends on HADOOP-9918. 

> Separate out RM services into "Always On" and "Active"
> --
>
> Key: YARN-1098
> URL: https://issues.apache.org/jira/browse/YARN-1098
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Affects Versions: 2.1.0-beta
>Reporter: Karthik Kambatla
>Assignee: Karthik Kambatla
>  Labels: ha
> Attachments: yarn-1098-1.patch, yarn-1098-approach.patch, 
> yarn-1098-approach.patch
>
>
> From discussion on YARN-1027, it makes sense to separate out services that 
> are stateful and stateless. The stateless services can  run perennially 
> irrespective of whether the RM is in Active/Standby state, while the stateful 
> services need to  be started on transitionToActive() and completely shutdown 
> on transitionToStandby().
> The external-facing stateless services should respond to the client/AM/NM 
> requests depending on whether the RM is Active/Standby.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1127) reservation exchange and excess reservation is not working for capacity scheduler

2013-08-30 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755110#comment-13755110
 ] 

Bikas Saha commented on YARN-1127:
--

Then please clarify this in the description or comment. Otherwise it looked 
like an exact duplicate. So the purpose of this jira is to fix the following 
situation.
1) NM1 has 2048 capacity in total but only 512 is free. A reservation of 1024 
is placed on it
2) NM2 now reports 1024 free space. At this point, the above reservation should 
be removed from NM1 and container should be assigned to NM2.
Step 2 is not happening and this jira intends to fix it.

> reservation exchange and excess reservation is not working for capacity 
> scheduler
> -
>
> Key: YARN-1127
> URL: https://issues.apache.org/jira/browse/YARN-1127
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2. 
> * We support a notion where if say we have 5 nodes with 4 AM and all node 
> managers have 8GB each and AM 2 GB each. Each AM is requesting 8GB each. Now 
> to avoid deadlock AM will make an extra reservation. By doing this we would 
> never hit the deadlock situation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-696) Enable multiple states to to be specified in Resource Manager apps REST call

2013-08-30 Thread Zhijie Shen (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755162#comment-13755162
 ] 

Zhijie Shen commented on YARN-696:
--

The patch is almost good. Here's some minor comments:

1. To compare the strings with case ingored, is it better to use
{code}
+appStates.add(state.trim().toLowerCase());
{code}
and
{code}
+  if (appState.contains(rmapp.getState().toString().toLowerCase())) {
{code}
It may not have much performance gain given a small string collection, but it 
should make code more concise.
 
2. It would be more clear if the test can verify one app is ACCEPTED and the 
other is KILLED.
{code}
+assertTrue("no states equal to KILLED", 
+(array.getJSONObject(0).getString("state").equals("KILLED")) ||
+(array.getJSONObject(1).getString("state").equals("KILLED")));
+assertTrue("no states equal to ACCEPTED", 
+(array.getJSONObject(0).getString("state").equals("ACCEPTED")) ||
+(array.getJSONObject(1).getString("state").equals("ACCEPTED")));
{code}

> Enable multiple states to to be specified in Resource Manager apps REST call
> 
>
> Key: YARN-696
> URL: https://issues.apache.org/jira/browse/YARN-696
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Affects Versions: 2.0.4-alpha
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: YARN-696.diff, YARN-696.diff
>
>
> Within the YARN Resource Manager REST API the GET call which returns all 
> Applications can be filtered by a single State query parameter (http:// http address:port>/ws/v1/cluster/apps). 
> There are 8 possible states (New, Submitted, Accepted, Running, Finishing, 
> Finished, Failed, Killed), if no state parameter is specified all states are 
> returned, however if a sub-set of states is required then multiple REST calls 
> are required (max. of 7).
> The proposal is to be able to specify multiple states in a single REST call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755063#comment-13755063
 ] 

Hadoop QA commented on YARN-1065:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600840/YARN-1065.2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1809//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1809//console

This message is automatically generated.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1129) Job hungs when any node is blacklisted after RMrestart

2013-08-30 Thread yeshavora (JIRA)
yeshavora created YARN-1129:
---

 Summary: Job hungs when any node is blacklisted after RMrestart
 Key: YARN-1129
 URL: https://issues.apache.org/jira/browse/YARN-1129
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: yeshavora


When RM restarted, if during restart one NM went bad (bad disk), NM got 
blacklisted by AM and RM keeps giving the containers on the same node even 
though AM doesn't want it there.

Need to change AM to specifically blacklist node in the RM requests.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-896) Roll up for long lived YARN

2013-08-30 Thread Jason Lowe (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-896?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754812#comment-13754812
 ] 

Jason Lowe commented on YARN-896:
-

bq. Chris, feel free to file a JIRA for rolling of stdout and stderr and we can 
look into what it will take to support that properly.

[~ste...@apache.org] recently filed YARN-1104 as a subtask of this JIRA which 
covers the NM rolling stdout/stderr.  We can transmute that JIRA into whatever 
ends up rolling the logs if it's not the NM.

> Roll up for long lived YARN
> ---
>
> Key: YARN-896
> URL: https://issues.apache.org/jira/browse/YARN-896
> Project: Hadoop YARN
>  Issue Type: New Feature
>Reporter: Robert Joseph Evans
>
> YARN is intended to be general purpose, but it is missing some features to be 
> able to truly support long lived applications and long lived containers.
> This ticket is intended to
>  # discuss what is needed to support long lived processes
>  # track the resulting JIRA.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1129) Job hungs when any node is blacklisted after RMrestart

2013-08-30 Thread Zhijie Shen (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijie Shen updated YARN-1129:
--

Assignee: Zhijie Shen

> Job hungs when any node is blacklisted after RMrestart
> --
>
> Key: YARN-1129
> URL: https://issues.apache.org/jira/browse/YARN-1129
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: yeshavora
>Assignee: Zhijie Shen
>
> When RM restarted, if during restart one NM went bad (bad disk), NM got 
> blacklisted by AM and RM keeps giving the containers on the same node even 
> though AM doesn't want it there.
> Need to change AM to specifically blacklist node in the RM requests.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-957) Capacity Scheduler tries to reserve the memory more than what node manager reports.

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755115#comment-13755115
 ] 

Hadoop QA commented on YARN-957:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12600849/YARN-957-20130830.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1810//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1810//console

This message is automatically generated.

> Capacity Scheduler tries to reserve the memory more than what node manager 
> reports.
> ---
>
> Key: YARN-957
> URL: https://issues.apache.org/jira/browse/YARN-957
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
> Attachments: YARN-957-20130730.1.patch, YARN-957-20130730.2.patch, 
> YARN-957-20130730.3.patch, YARN-957-20130731.1.patch, 
> YARN-957-20130830.1.patch
>
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * It should not try to reserve memory on a node manager which is never going 
> to give requested memory. i.e. Current max capability of node manager is 
> 1024MB but 2048MB is reserved on it. But it still does that.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1127) reservation exchange and excess reservation is not working for capacity scheduler

2013-08-30 Thread Omkar Vinit Joshi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755103#comment-13755103
 ] 

Omkar Vinit Joshi commented on YARN-1127:
-

No as per Arun I am separating out issues which are causing this failure.
* YARN-957 :- Fix if container is getting reserved on a node manager which 
exceeds its memory.
* this jira :- Ideally the switch should have taken place from one to other 
node manager if another node manager has sufficient memory. However that did 
not happen. This must have occurred either because excess reservation did not 
work or reservation exchange did not occur. We need to find the root cause  and 
fix this.

> reservation exchange and excess reservation is not working for capacity 
> scheduler
> -
>
> Key: YARN-1127
> URL: https://issues.apache.org/jira/browse/YARN-1127
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2. 
> * We support a notion where if say we have 5 nodes with 4 AM and all node 
> managers have 8GB each and AM 2 GB each. Each AM is requesting 8GB each. Now 
> to avoid deadlock AM will make an extra reservation. By doing this we would 
> never hit the deadlock situation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (YARN-1128) FifoPolicy.computeShares throws NPE on empty list of Schedulables

2013-08-30 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1128?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla reassigned YARN-1128:
--

Assignee: Karthik Kambatla

> FifoPolicy.computeShares throws NPE on empty list of Schedulables
> -
>
> Key: YARN-1128
> URL: https://issues.apache.org/jira/browse/YARN-1128
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: scheduler
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Karthik Kambatla
>
> FifoPolicy gives all of a queue's share to the earliest-scheduled application.
> {code}
> Schedulable earliest = null;
> for (Schedulable schedulable : schedulables) {
>   if (earliest == null ||
>   schedulable.getStartTime() < earliest.getStartTime()) {
> earliest = schedulable;
>   }
> }
> earliest.setFairShare(Resources.clone(totalResources));
> {code}
> If the queue has no schedulables in it, earliest will be left null, leading 
> to an NPE on the last line.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-696) Enable multiple states to to be specified in Resource Manager apps REST call

2013-08-30 Thread Trevor Lorimer (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-696?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Trevor Lorimer updated YARN-696:


Attachment: YARN-696.diff

Okay i'll take your word for it, uploading latest diff.

The unit tests use 2 different applications with states of ACCEPTED and KILLED. 

> Enable multiple states to to be specified in Resource Manager apps REST call
> 
>
> Key: YARN-696
> URL: https://issues.apache.org/jira/browse/YARN-696
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Affects Versions: 2.0.4-alpha
>Reporter: Trevor Lorimer
>Assignee: Trevor Lorimer
> Attachments: YARN-696.diff, YARN-696.diff
>
>
> Within the YARN Resource Manager REST API the GET call which returns all 
> Applications can be filtered by a single State query parameter (http:// http address:port>/ws/v1/cluster/apps). 
> There are 8 possible states (New, Submitted, Accepted, Running, Finishing, 
> Finished, Failed, Killed), if no state parameter is specified all states are 
> returned, however if a sub-set of states is required then multiple REST calls 
> are required (max. of 7).
> The proposal is to be able to specify multiple states in a single REST call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1127) reservation exchange and excess reservation is not working for capacity scheduler

2013-08-30 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755095#comment-13755095
 ] 

Bikas Saha commented on YARN-1127:
--

How is this different from YARN-957

> reservation exchange and excess reservation is not working for capacity 
> scheduler
> -
>
> Key: YARN-1127
> URL: https://issues.apache.org/jira/browse/YARN-1127
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2. 
> * We support a notion where if say we have 5 nodes with 4 AM and all node 
> managers have 8GB each and AM 2 GB each. Each AM is requesting 8GB each. Now 
> to avoid deadlock AM will make an extra reservation. By doing this we would 
> never hit the deadlock situation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1127) reservation exchange and excess reservation is not working for capacity scheduler

2013-08-30 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755090#comment-13755090
 ] 

Bikas Saha commented on YARN-1127:
--

Isnt this similar to a jira opened by you already? The issue being that the 
scheduler puts a reservation on a node whose total capacity is smaller than the 
reservation resource size. In this case, nm1 has capacity=1024 but the 
scheduler is putting a reservation of 2048 on it and that can never be 
satisfied. So it does not make sense to make that reservation at all.

> reservation exchange and excess reservation is not working for capacity 
> scheduler
> -
>
> Key: YARN-1127
> URL: https://issues.apache.org/jira/browse/YARN-1127
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.1.1-beta
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2. 
> * We support a notion where if say we have 5 nodes with 4 AM and all node 
> managers have 8GB each and AM 2 GB each. Each AM is requesting 8GB each. Now 
> to avoid deadlock AM will make an extra reservation. By doing this we would 
> never hit the deadlock situation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-957) Capacity Scheduler tries to reserve the memory more than what node manager reports.

2013-08-30 Thread Omkar Vinit Joshi (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-957?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Omkar Vinit Joshi updated YARN-957:
---

Attachment: YARN-957-20130830.1.patch

> Capacity Scheduler tries to reserve the memory more than what node manager 
> reports.
> ---
>
> Key: YARN-957
> URL: https://issues.apache.org/jira/browse/YARN-957
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
> Attachments: YARN-957-20130730.1.patch, YARN-957-20130730.2.patch, 
> YARN-957-20130730.3.patch, YARN-957-20130731.1.patch, 
> YARN-957-20130830.1.patch
>
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * It should not try to reserve memory on a node manager which is never going 
> to give requested memory. i.e. Current max capability of node manager is 
> 1024MB but 2048MB is reserved on it. But it still does that.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-957) Capacity Scheduler tries to reserve the memory more than what node manager reports.

2013-08-30 Thread Omkar Vinit Joshi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755089#comment-13755089
 ] 

Omkar Vinit Joshi commented on YARN-957:


uploading a patch which only fixes excess memory reservation issue.

> Capacity Scheduler tries to reserve the memory more than what node manager 
> reports.
> ---
>
> Key: YARN-957
> URL: https://issues.apache.org/jira/browse/YARN-957
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Omkar Vinit Joshi
>Assignee: Omkar Vinit Joshi
>Priority: Blocker
> Attachments: YARN-957-20130730.1.patch, YARN-957-20130730.2.patch, 
> YARN-957-20130730.3.patch, YARN-957-20130731.1.patch
>
>
> I have 2 node managers.
> * one with 1024 MB memory.(nm1)
> * second with 2048 MB memory.(nm2)
> I am submitting simple map reduce application with 1 mapper and one reducer 
> with 1024mb each. The steps to reproduce this are
> * stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
> heartbeat doesn't reach RM first).
> * now submit application. As soon as it receives first node's (nm1) heartbeat 
> it will try to reserve memory for AM-container (2048MB). However it has only 
> 1024MB of memory.
> * now start nm2 with 2048 MB memory.
> It hangs forever... Ideally this has two potential issues.
> * It should not try to reserve memory on a node manager which is never going 
> to give requested memory. i.e. Current max capability of node manager is 
> 1024MB but 2048MB is reserved on it. But it still does that.
> * Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
> memory. In this case if the original request was made without any locality 
> then scheduler should unreserve memory on nm1 and allocate requested 2048MB 
> container on nm2.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755088#comment-13755088
 ] 

Bikas Saha commented on YARN-1065:
--

There seems to be some misunderstanding here. The NM already provides service 
metadata in startContainerResponse to the AM.
The main issue is that when the NM actually launches the process of that 
container then that process has no way to find information about the auxillary 
services mentioned in the startContainerRequest for that container.
So the desired solution is the following
1) StartContainerRequest sends request for auxillary services to NM (this 
already happens today)
2) NM sends back auxillary service metadata (say Map) in the 
StartContainerResponse (this happens today)
3) When NM launches the process for this container then it should put the 
service metadata (Map) in the container environment. Eg. add 
to environment key=FOO_NM_SERVICE_KEY with value=base64String(ByteBuffer).
4) There should be some helper API to enable the process in the container to 
retrieve those keys and value. Process uses api, say, 
Helper.getServiceMetadata(FOO) which looks up FOO_NM_SERVICE_KEY in the env and 
converts the base64String value back to ByteBuffer. NM could also user 
Helper.set(FOO, ByteBuffer, env) that sets key=FOO_NM_SERVICE_KEY and 
base64String(ByteBuffer) into the env. Using helper API will keep both sides 
consistent.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1104) NMs to support rolling logs of stdout & stderr

2013-08-30 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755087#comment-13755087
 ] 

Steve Loughran commented on YARN-1104:
--

Currently most output setups are being done in the command line itself with a 
>/out.txt 2>/err.txt

we should allow apps to ask for the NM set up the logging, which could then be 
a direct capture from the process streams and rolling logging into the 
directories, maybe even allow for other options. Routing it via log4j isn't 
ideal because it does tend to add extra noise:
{code}
d=0}
2013-08-29 17:07:40,594 [AMRM Callback Handler Thread] DEBUG 
yarn.appmaster.HoyaAppMaster (HoyaAppMaster.java:getProgress(1280)) - 
Heartbeat, percentage =0.5
2013-08-29 17:07:40,900 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:40 DEBUG client.ClientScanner: 
Finished region={ENCODED => 92bbefc5114a871444d6665dbddd7fda, NAME => 
'hbase:namesp
2013-08-29 17:07:40,900 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 
ace,,1377821255366.92bbefc5114a871444d6665dbddd7fda.', STARTKEY => '', ENDKEY 
=> ''}
2013-08-29 17:07:41,101 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:40 INFO 
zookeeper.RecoverableZooKeeper: Node 
/yarnapps_hoya_stevel_TestClusterFlex2DownTo1/namespace/default alrea
2013-08-29 17:07:41,101 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - dy exists and this is not a retry
2013-08-29 17:07:41,101 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:41 DEBUG 
hbase.ZKNamespaceManager: Updating namespace cache from node default with data: 
\x0A\x07default
2013-08-29 17:07:41,101 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:41 DEBUG 
hbase.ZKNamespaceManager: Updating namespace cache from node hbase with data: 
\x0A\x05hbase
2013-08-29 17:07:41,429 [IPC Server handler 1 on 52037] INFO  
yarn.appmaster.HoyaAppMaster (HoyaAppMaster.java:stopCluster(1306)) - 
HoyaAppMasterApi.stopCluster()
2013-08-29 17:07:41,429 [IPC Server handler 1 on 52037] INFO  
yarn.appmaster.HoyaAppMaster (HoyaAppMaster.java:signalAMComplete(709)) - 
Triggering shutdown of the AM: stopCluster() invoked
2013-08-29 17:07:41,502 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:41 INFO 
zookeeper.RecoverableZooKeeper: Node 
/yarnapps_hoya_stevel_TestClusterFlex2DownTo1/namespace/hbase already
2013-08-29 17:07:41,503 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) -  exists and this is not a retry
2013-08-29 17:07:41,597 [AMRM Callback Handler Thread] DEBUG 
yarn.appmaster.HoyaAppMaster (HoyaAppMaster.java:getProgress(1280)) - 
Heartbeat, percentage =0.5
2013-08-29 17:07:41,904 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - 13/08/29 17:07:41 DEBUG client.ClientScanner: 
Finished region={ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => 
'', EN
2013-08-29 17:07:41,904 [Thread-12] WARN  hoya.exec.RunLongLivedApp 
(RunLongLivedApp.java:run(351)) - DKEY => ''}

{code}

Perhaps we could have a log4j formatter that does nothing at all -just prints 
the output direct -then route it to a specific log roll just for the outputs

> NMs to support rolling logs of stdout & stderr
> --
>
> Key: YARN-1104
> URL: https://issues.apache.org/jira/browse/YARN-1104
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.1.0-beta
>Reporter: Steve Loughran
>
> Currently NMs stream the stdout and stderr streams of a container to a file. 
> For longer lived processes those files need to be rotated so that the log 
> doesn't overflow

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1128) FifoPolicy.computeShares throws NPE on empty list of Schedulables

2013-08-30 Thread Sandy Ryza (JIRA)
Sandy Ryza created YARN-1128:


 Summary: FifoPolicy.computeShares throws NPE on empty list of 
Schedulables
 Key: YARN-1128
 URL: https://issues.apache.org/jira/browse/YARN-1128
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.1.0-beta
Reporter: Sandy Ryza


FifoPolicy gives all of a queue's share to the earliest-scheduled application.

{code}
Schedulable earliest = null;
for (Schedulable schedulable : schedulables) {
  if (earliest == null ||
  schedulable.getStartTime() < earliest.getStartTime()) {
earliest = schedulable;
  }
}
earliest.setFairShare(Resources.clone(totalResources));
{code}

If the queue has no schedulables in it, earliest will be left null, leading to 
an NPE on the last line.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1120) Make ApplicationConstants.Environment.USER definition OS neutral

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1120?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755055#comment-13755055
 ] 

Hadoop QA commented on YARN-1120:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600495/YARN-1120.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1808//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1808//console

This message is automatically generated.

> Make ApplicationConstants.Environment.USER definition OS neutral
> 
>
> Key: YARN-1120
> URL: https://issues.apache.org/jira/browse/YARN-1120
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Chuan Liu
>Assignee: Chuan Liu
>Priority: Minor
> Attachments: YARN-1120.patch
>
>
> In YARN-557, we added some code to make {{ 
> ApplicationConstants.Environment.USER}} has OS-specific definition in order 
> to fix the unit test TestUnmanagedAMLauncher. In YARN-571, the relevant test 
> code was corrected. In YARN-602, we actually will explicitly set the 
> environment variables for the child containers. With these changes, I think 
> we can revert the YARN-557 change to make {{ 
> ApplicationConstants.Environment.USER}} OS neutral. The main benefit is that 
> we can use the same method over the Enum constants. This should also fix the 
> TestContainerLaunch#testContainerEnvVariables failure on Windows. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-707) Add user info in the YARN ClientToken

2013-08-30 Thread Jason Lowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe updated YARN-707:


Attachment: YARN-707-20130830.branch-0.23.txt

Patch for branch-0.23.  [~vinodkv] could you take a peek at it?

All of the tests in mapreduce-client-app, yarn-common, and 
yarn-server-resourcemanager  pass after this change, and I manually tested it 
on a secure cluster with the branch-0.23 patch from MAPREDUCE-5475 to verify 
ACLs can be implemented properly.

> Add user info in the YARN ClientToken
> -
>
> Key: YARN-707
> URL: https://issues.apache.org/jira/browse/YARN-707
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Jason Lowe
>Priority: Blocker
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: YARN-707-20130822.txt, YARN-707-20130827.txt, 
> YARN-707-20130828-2.txt, YARN-707-20130828.txt, YARN-707-20130829.txt, 
> YARN-707-20130830.branch-0.23.txt
>
>
> If user info is present in the client token then it can be used to do limited 
> authz in the AM.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1127) reservation exchange and excess reservation is not working for capacity scheduler

2013-08-30 Thread Omkar Vinit Joshi (JIRA)
Omkar Vinit Joshi created YARN-1127:
---

 Summary: reservation exchange and excess reservation is not 
working for capacity scheduler
 Key: YARN-1127
 URL: https://issues.apache.org/jira/browse/YARN-1127
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.1.1-beta
Reporter: Omkar Vinit Joshi
Assignee: Omkar Vinit Joshi
Priority: Blocker


I have 2 node managers.
* one with 1024 MB memory.(nm1)
* second with 2048 MB memory.(nm2)
I am submitting simple map reduce application with 1 mapper and one reducer 
with 1024mb each. The steps to reproduce this are
* stop nm2 with 2048MB memory.( This I am doing to make sure that this node's 
heartbeat doesn't reach RM first).
* now submit application. As soon as it receives first node's (nm1) heartbeat 
it will try to reserve memory for AM-container (2048MB). However it has only 
1024MB of memory.
* now start nm2 with 2048 MB memory.

It hangs forever... Ideally this has two potential issues.

* Say 2048MB is reserved on nm1 but nm2 comes back with 2048MB available 
memory. In this case if the original request was made without any locality then 
scheduler should unreserve memory on nm1 and allocate requested 2048MB 
container on nm2. 
* We support a notion where if say we have 5 nodes with 4 AM and all node 
managers have 8GB each and AM 2 GB each. Each AM is requesting 8GB each. Now to 
avoid deadlock AM will make an extra reservation. By doing this we would never 
hit the deadlock situation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1065:


Attachment: YARN-1065.2.patch

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13755048#comment-13755048
 ] 

Xuan Gong commented on YARN-1065:
-

[~bikassaha]
Please ignore the YARN-1065.1.patch, please take a look at YARN-1065.2.patch.

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch, YARN-1065.2.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1065) NM should provide AuxillaryService data to the container

2013-08-30 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1065:


Attachment: YARN-1065.1.patch

> NM should provide AuxillaryService data to the container
> 
>
> Key: YARN-1065
> URL: https://issues.apache.org/jira/browse/YARN-1065
> Project: Hadoop YARN
>  Issue Type: Task
>Reporter: Bikas Saha
>Assignee: Xuan Gong
> Attachments: YARN-1065.1.patch
>
>
> Start container returns auxillary service data to the AM but does not provide 
> the same information to the task itself. It could add that information to the 
> container env with key=service_name and value=service_data. This allows the 
> container to start using the service without having to depend on the AM to 
> send the info to it indirectly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-896) Roll up for long lived YARN

2013-08-30 Thread Robert Joseph Evans (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-896?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754787#comment-13754787
 ] 

Robert Joseph Evans commented on YARN-896:
--

I agree that providing a good way handle stdout and stderr is important. I 
don't know if I want the NM to be doing this for us though, but that is an 
implementation detail that we can talk about on the follow up JIRA.  Chris, 
feel free to file a JIRA for rolling of stdout and stderr and we can look into 
what it will take to support that properly.

> Roll up for long lived YARN
> ---
>
> Key: YARN-896
> URL: https://issues.apache.org/jira/browse/YARN-896
> Project: Hadoop YARN
>  Issue Type: New Feature
>Reporter: Robert Joseph Evans
>
> YARN is intended to be general purpose, but it is missing some features to be 
> able to truly support long lived applications and long lived containers.
> This ticket is intended to
>  # discuss what is needed to support long lived processes
>  # track the resulting JIRA.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-707) Add user info in the YARN ClientToken

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754741#comment-13754741
 ] 

Hudson commented on YARN-707:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1534 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1534/])
YARN-707. Added user information also in the YARN ClientToken so that AMs can 
implement authorization based on incoming users. Contributed by Jason Lowe. 
(vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518868)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenIdentifier.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenSecretManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMAppImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttempt.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttemptImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/ClientToAMTokenSecretManagerInRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationsmanager/MockAsm.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/MockRMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/TestRMAppTransitions.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java


> Add user info in the YARN ClientToken
> -
>
> Key: YARN-707
> URL: https://issues.apache.org/jira/browse/YARN-707
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Jason Lowe
>Priority: Blocker
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: YARN-707-20130822.txt, YARN-707-20130827.txt, 
> YARN-707-20130828-2.txt, YARN-707-20130828.txt, YARN-707-20130829.txt
>
>
> If user info is present in the client token then it can be used to do limited 
> authz in the AM.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1080) Improve help message for $ yarn logs

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754747#comment-13754747
 ] 

Hudson commented on YARN-1080:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1534 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1534/])
YARN-1080. Improved help message for "yarn logs" command. Contributed by Xuan 
Gong. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518731)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/LogDumper.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/logaggregation/TestLogDumper.java


> Improve help message for $ yarn logs
> 
>
> Key: YARN-1080
> URL: https://issues.apache.org/jira/browse/YARN-1080
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1080.1.patch, YARN-1080.2.patch
>
>
> There are 2 parts I am proposing in this jira. They can be fixed together in 
> one patch.
> 1. Standardize help message for required parameter of $ yarn logs
> YARN CLI has a command "logs" ($ yarn logs). The command always requires a 
> parameter of "-applicationId ". However, help message of the command 
> does not make it clear. It lists -applicationId as optional parameter. If I 
> don't set it, YARN CLI will complain this is missing. It is better to use 
> standard required notation used in other Linux command for help message. Any 
> user familiar to the command can understand that this parameter is needed 
> more easily.
> {code:title=current help message}
> -bash-4.1$ yarn logs
> usage: general options are:
>  -applicationIdApplicationId (required)
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> {code:title=proposed help message}
> -bash-4.1$ yarn logs
> usage: yarn logs -applicationId  [OPTIONS]
> general options are:
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> 2. Add description for help command. As far as I know, a user cannot get logs 
> for running job. Since I spent some time trying to get logs of running 
> applications, it should be nice to say this in command description.
> {code:title=proposed help}
> Retrieve logs for completed/killed YARN application
> usage: general options are...
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-707) Add user info in the YARN ClientToken

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754665#comment-13754665
 ] 

Hudson commented on YARN-707:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1507 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1507/])
YARN-707. Added user information also in the YARN ClientToken so that AMs can 
implement authorization based on incoming users. Contributed by Jason Lowe. 
(vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518868)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenIdentifier.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenSecretManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMAppImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttempt.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttemptImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/ClientToAMTokenSecretManagerInRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationsmanager/MockAsm.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/MockRMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/TestRMAppTransitions.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java


> Add user info in the YARN ClientToken
> -
>
> Key: YARN-707
> URL: https://issues.apache.org/jira/browse/YARN-707
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Jason Lowe
>Priority: Blocker
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: YARN-707-20130822.txt, YARN-707-20130827.txt, 
> YARN-707-20130828-2.txt, YARN-707-20130828.txt, YARN-707-20130829.txt
>
>
> If user info is present in the client token then it can be used to do limited 
> authz in the AM.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1080) Improve help message for $ yarn logs

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754671#comment-13754671
 ] 

Hudson commented on YARN-1080:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk #1507 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1507/])
YARN-1080. Improved help message for "yarn logs" command. Contributed by Xuan 
Gong. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518731)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/LogDumper.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/logaggregation/TestLogDumper.java


> Improve help message for $ yarn logs
> 
>
> Key: YARN-1080
> URL: https://issues.apache.org/jira/browse/YARN-1080
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1080.1.patch, YARN-1080.2.patch
>
>
> There are 2 parts I am proposing in this jira. They can be fixed together in 
> one patch.
> 1. Standardize help message for required parameter of $ yarn logs
> YARN CLI has a command "logs" ($ yarn logs). The command always requires a 
> parameter of "-applicationId ". However, help message of the command 
> does not make it clear. It lists -applicationId as optional parameter. If I 
> don't set it, YARN CLI will complain this is missing. It is better to use 
> standard required notation used in other Linux command for help message. Any 
> user familiar to the command can understand that this parameter is needed 
> more easily.
> {code:title=current help message}
> -bash-4.1$ yarn logs
> usage: general options are:
>  -applicationIdApplicationId (required)
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> {code:title=proposed help message}
> -bash-4.1$ yarn logs
> usage: yarn logs -applicationId  [OPTIONS]
> general options are:
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> 2. Add description for help command. As far as I know, a user cannot get logs 
> for running job. Since I spent some time trying to get logs of running 
> applications, it should be nice to say this in command description.
> {code:title=proposed help}
> Retrieve logs for completed/killed YARN application
> usage: general options are...
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1080) Improve help message for $ yarn logs

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754586#comment-13754586
 ] 

Hudson commented on YARN-1080:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #317 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/317/])
YARN-1080. Improved help message for "yarn logs" command. Contributed by Xuan 
Gong. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518731)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/logaggregation/LogDumper.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/logaggregation/TestLogDumper.java


> Improve help message for $ yarn logs
> 
>
> Key: YARN-1080
> URL: https://issues.apache.org/jira/browse/YARN-1080
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1080.1.patch, YARN-1080.2.patch
>
>
> There are 2 parts I am proposing in this jira. They can be fixed together in 
> one patch.
> 1. Standardize help message for required parameter of $ yarn logs
> YARN CLI has a command "logs" ($ yarn logs). The command always requires a 
> parameter of "-applicationId ". However, help message of the command 
> does not make it clear. It lists -applicationId as optional parameter. If I 
> don't set it, YARN CLI will complain this is missing. It is better to use 
> standard required notation used in other Linux command for help message. Any 
> user familiar to the command can understand that this parameter is needed 
> more easily.
> {code:title=current help message}
> -bash-4.1$ yarn logs
> usage: general options are:
>  -applicationIdApplicationId (required)
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> {code:title=proposed help message}
> -bash-4.1$ yarn logs
> usage: yarn logs -applicationId  [OPTIONS]
> general options are:
>  -appOwner AppOwner (assumed to be current user if not
> specified)
>  -containerId  ContainerId (must be specified if node address is
> specified)
>  -nodeAddress  NodeAddress in the format nodename:port (must be
> specified if container id is specified)
> {code}
> 2. Add description for help command. As far as I know, a user cannot get logs 
> for running job. Since I spent some time trying to get logs of running 
> applications, it should be nice to say this in command description.
> {code:title=proposed help}
> Retrieve logs for completed/killed YARN application
> usage: general options are...
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-707) Add user info in the YARN ClientToken

2013-08-30 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754580#comment-13754580
 ] 

Hudson commented on YARN-707:
-

SUCCESS: Integrated in Hadoop-Yarn-trunk #317 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/317/])
YARN-707. Added user information also in the YARN ClientToken so that AMs can 
implement authorization based on incoming users. Contributed by Jason Lowe. 
(vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1518868)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenIdentifier.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/client/ClientToAMTokenSecretManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/RMAppImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttempt.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/attempt/RMAppAttemptImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/security/ClientToAMTokenSecretManagerInRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/applicationsmanager/MockAsm.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/MockRMApp.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/rmapp/TestRMAppTransitions.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java


> Add user info in the YARN ClientToken
> -
>
> Key: YARN-707
> URL: https://issues.apache.org/jira/browse/YARN-707
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Jason Lowe
>Priority: Blocker
> Fix For: 3.0.0, 2.1.1-beta
>
> Attachments: YARN-707-20130822.txt, YARN-707-20130827.txt, 
> YARN-707-20130828-2.txt, YARN-707-20130828.txt, YARN-707-20130829.txt
>
>
> If user info is present in the client token then it can be used to do limited 
> authz in the AM.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-879) Fix tests w.r.t o.a.h.y.server.resourcemanager.Application

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-879?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754505#comment-13754505
 ] 

Hadoop QA commented on YARN-879:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600751/YARN-879-v3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 5 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1806//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1806//console

This message is automatically generated.

> Fix tests w.r.t o.a.h.y.server.resourcemanager.Application
> --
>
> Key: YARN-879
> URL: https://issues.apache.org/jira/browse/YARN-879
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.1.0-beta
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: YARN-879.patch, YARN-879-v2.patch, YARN-879-v3.patch
>
>
> getResources() will return a list of containers that allocated by RM. 
> However, it is now return null directly. The worse thing is: if LOG.debug is 
> enabled, then it will definitely cause NPE exception.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-771) AMRMClient support for resource blacklisting

2013-08-30 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13754494#comment-13754494
 ] 

Hadoop QA commented on YARN-771:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600749/YARN-771-v4.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/1805//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/1805//console

This message is automatically generated.

> AMRMClient  support for resource blacklisting
> -
>
> Key: YARN-771
> URL: https://issues.apache.org/jira/browse/YARN-771
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Bikas Saha
>Assignee: Junping Du
> Attachments: YARN-771-v1.0.patch, YARN-771-v2.patch, 
> YARN-771-v3.patch, YARN-771-v4.patch
>
>
> After YARN-750 AMRMClient should support blacklisting via the new YARN API's

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-879) Fix tests w.r.t o.a.h.y.server.resourcemanager.Application

2013-08-30 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-879?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated YARN-879:


Attachment: YARN-879-v3.patch

Sync up patch with latest trunk branch. Can someone take a review on this? Thx!

> Fix tests w.r.t o.a.h.y.server.resourcemanager.Application
> --
>
> Key: YARN-879
> URL: https://issues.apache.org/jira/browse/YARN-879
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.1.0-beta
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: YARN-879.patch, YARN-879-v2.patch, YARN-879-v3.patch
>
>
> getResources() will return a list of containers that allocated by RM. 
> However, it is now return null directly. The worse thing is: if LOG.debug is 
> enabled, then it will definitely cause NPE exception.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-771) AMRMClient support for resource blacklisting

2013-08-30 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated YARN-771:


Attachment: YARN-771-v4.patch

> AMRMClient  support for resource blacklisting
> -
>
> Key: YARN-771
> URL: https://issues.apache.org/jira/browse/YARN-771
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Bikas Saha
>Assignee: Junping Du
> Attachments: YARN-771-v1.0.patch, YARN-771-v2.patch, 
> YARN-771-v3.patch, YARN-771-v4.patch
>
>
> After YARN-750 AMRMClient should support blacklisting via the new YARN API's

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira