[jira] [Created] (YARN-2013) The diagnostics is always the ExitCodeException stack when the container crashes

2014-05-01 Thread Zhijie Shen (JIRA)
Zhijie Shen created YARN-2013:
-

 Summary: The diagnostics is always the ExitCodeException stack 
when the container crashes
 Key: YARN-2013
 URL: https://issues.apache.org/jira/browse/YARN-2013
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Reporter: Zhijie Shen


When a container crashes, ExitCodeException will be thrown from Shell. 
Default/LinuxContainerExecutor captures the exception, put the exception stack 
into the diagnostic. Therefore, the exception stack is always the same. 
{code}
String diagnostics = Exception from container-launch: \n
+ StringUtils.stringifyException(e) + \n + shExec.getOutput();
container.handle(new ContainerDiagnosticsUpdateEvent(containerId,
diagnostics));
{code}

In addition, it seems that the exception always has a empty message as there's 
no message from stderr. Hence the diagnostics is not of much use for users to 
analyze the reason of container crash.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-2013) The diagnostics is always the ExitCodeException stack when the container crashes

2014-05-01 Thread Tsuyoshi OZAWA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13986560#comment-13986560
 ] 

Tsuyoshi OZAWA commented on YARN-2013:
--

+1(non-binding), in fact I often faces the problem in Tez project. This 
behavior make it difficult to debug YARN Apps.

TestContainerLaunch#testContainerLaunchStdoutAndStderrDiagnostics() confirms 
that Exception.getMessages() returns the stderr's message. We should include 
the message into the diagnostics.

 The diagnostics is always the ExitCodeException stack when the container 
 crashes
 

 Key: YARN-2013
 URL: https://issues.apache.org/jira/browse/YARN-2013
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Reporter: Zhijie Shen

 When a container crashes, ExitCodeException will be thrown from Shell. 
 Default/LinuxContainerExecutor captures the exception, put the exception 
 stack into the diagnostic. Therefore, the exception stack is always the same. 
 {code}
 String diagnostics = Exception from container-launch: \n
 + StringUtils.stringifyException(e) + \n + shExec.getOutput();
 container.handle(new ContainerDiagnosticsUpdateEvent(containerId,
 diagnostics));
 {code}
 In addition, it seems that the exception always has a empty message as 
 there's no message from stderr. Hence the diagnostics is not of much use for 
 users to analyze the reason of container crash.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (YARN-2013) The diagnostics is always the ExitCodeException stack when the container crashes

2014-05-01 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2013?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA reassigned YARN-2013:


Assignee: Tsuyoshi OZAWA

 The diagnostics is always the ExitCodeException stack when the container 
 crashes
 

 Key: YARN-2013
 URL: https://issues.apache.org/jira/browse/YARN-2013
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Reporter: Zhijie Shen
Assignee: Tsuyoshi OZAWA

 When a container crashes, ExitCodeException will be thrown from Shell. 
 Default/LinuxContainerExecutor captures the exception, put the exception 
 stack into the diagnostic. Therefore, the exception stack is always the same. 
 {code}
 String diagnostics = Exception from container-launch: \n
 + StringUtils.stringifyException(e) + \n + shExec.getOutput();
 container.handle(new ContainerDiagnosticsUpdateEvent(containerId,
 diagnostics));
 {code}
 In addition, it seems that the exception always has a empty message as 
 there's no message from stderr. Hence the diagnostics is not of much use for 
 users to analyze the reason of container crash.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-2013) The diagnostics is always the ExitCodeException stack when the container crashes

2014-05-01 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2013?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-2013:
-

Attachment: YARN-2013.1.patch

Changed to include stderr output in the diagnostics.

 The diagnostics is always the ExitCodeException stack when the container 
 crashes
 

 Key: YARN-2013
 URL: https://issues.apache.org/jira/browse/YARN-2013
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Reporter: Zhijie Shen
Assignee: Tsuyoshi OZAWA
 Attachments: YARN-2013.1.patch


 When a container crashes, ExitCodeException will be thrown from Shell. 
 Default/LinuxContainerExecutor captures the exception, put the exception 
 stack into the diagnostic. Therefore, the exception stack is always the same. 
 {code}
 String diagnostics = Exception from container-launch: \n
 + StringUtils.stringifyException(e) + \n + shExec.getOutput();
 container.handle(new ContainerDiagnosticsUpdateEvent(containerId,
 diagnostics));
 {code}
 In addition, it seems that the exception always has a empty message as 
 there's no message from stderr. Hence the diagnostics is not of much use for 
 users to analyze the reason of container crash.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1342) Recover container tokens upon nodemanager restart

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1342?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13986605#comment-13986605
 ] 

Hadoop QA commented on YARN-1342:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12642738/YARN-1342v3-and-YARN-1987.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3670//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3670//console

This message is automatically generated.

 Recover container tokens upon nodemanager restart
 -

 Key: YARN-1342
 URL: https://issues.apache.org/jira/browse/YARN-1342
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: nodemanager
Affects Versions: 2.3.0
Reporter: Jason Lowe
Assignee: Jason Lowe
 Attachments: YARN-1342.patch, YARN-1342v2.patch, 
 YARN-1342v3-and-YARN-1987.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated YARN-1201:
-

Issue Type: Bug  (was: Test)

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13986644#comment-13986644
 ] 

Junping Du commented on YARN-1201:
--

Thanks [~leftnoteasy] for identifying the root cause and delivering a patch 
here. However, I don't think we should fix test here as test code itself is 
fine to me. We'd better to throw an exception with meaningful message rather 
than NPE when local hostname cannot be resolved. So I move this from test to 
bug.

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1339) Recover DeletionService state upon nodemanager restart

2014-05-01 Thread Jason Lowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1339?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe updated YARN-1339:
-

Attachment: YARN-1339v3-and-YARN-1987.patch

Updating the patch to address the DBException handling that was brought up in 
the MAPREDUCE-5652 review and applies here. Note that this now depends upon 
YARN-1987 as that provides the utility wrapper for the leveldb iterator to 
translate raw RuntimeException to the more helpful DBException so we can act 
accordingly when errors occur. The other notable change in the patch is 
renaming LevelDB to Leveldb for consistency with the existing 
LeveldbTimelineStore naming convention.

This latest patch includes the necessary pieces of YARN-1987 so it can compile 
and Jenkins can comment.


 Recover DeletionService state upon nodemanager restart
 --

 Key: YARN-1339
 URL: https://issues.apache.org/jira/browse/YARN-1339
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: nodemanager
Affects Versions: 2.3.0
Reporter: Jason Lowe
Assignee: Jason Lowe
 Attachments: YARN-1339.patch, YARN-1339v2.patch, 
 YARN-1339v3-and-YARN-1987.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy

2014-05-01 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13986679#comment-13986679
 ] 

Sunil G commented on YARN-2009:
---

As per YARN-1963, jobs can be submitted in a queue with priority. 
In that scenario, when a queue is identified to preempt few containers among 
its applications, job priority also can be considered.


 Priority support for preemption in ProportionalCapacityPreemptionPolicy
 ---

 Key: YARN-2009
 URL: https://issues.apache.org/jira/browse/YARN-2009
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: capacityscheduler
Reporter: Devaraj K

 While preempting containers based on the queue ideal assignment, we may need 
 to consider preempting the low priority application containers first.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1864) Fair Scheduler Dynamic Hierarchical User Queues

2014-05-01 Thread Ashwin Shankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1864?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashwin Shankar updated YARN-1864:
-

Attachment: YARN-1864-v4.txt

 Fair Scheduler Dynamic Hierarchical User Queues
 ---

 Key: YARN-1864
 URL: https://issues.apache.org/jira/browse/YARN-1864
 Project: Hadoop YARN
  Issue Type: New Feature
  Components: scheduler
Reporter: Ashwin Shankar
  Labels: scheduler
 Attachments: YARN-1864-v1.txt, YARN-1864-v2.txt, YARN-1864-v3.txt, 
 YARN-1864-v4.txt


 In Fair Scheduler, we want to be able to create user queues under any parent 
 queue in the hierarchy. For eg. Say user1 submits a job to a parent queue 
 called root.allUserQueues, we want be able to create a new queue called 
 root.allUserQueues.user1 and run user1's job in it.Any further jobs submitted 
 by this user to root.allUserQueues will be run in this newly created 
 root.allUserQueues.user1.
 This is very similar to the 'user-as-default' feature in Fair Scheduler which 
 creates user queues under root queue. But we want the ability to create user 
 queues under ANY parent queue.
 Why do we want this ?
 1. Preemption : these dynamically created user queues can preempt each other 
 if its fair share is not met. So there is fairness among users.
 User queues can also preempt other non-user leaf queue as well if below fair 
 share.
 2. Allocation to user queues : we want all the user queries(adhoc) to consume 
 only a fraction of resources in the shared cluster. By creating this 
 feature,we could do that by giving a fair share to the parent user queue 
 which is then redistributed to all the dynamically created user queues.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1864) Fair Scheduler Dynamic Hierarchical User Queues

2014-05-01 Thread Ashwin Shankar (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1864?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987011#comment-13987011
 ] 

Ashwin Shankar commented on YARN-1864:
--

Hi [~sandyr] ,
bq. I think it's better to leave out case 4. The right behavior on it is fuzzy, 
and things are simpler if the results returned by QueuePlacementPolicy are only 
a function of the configuration.
Removed queueMgr and using configuredQueues instead.

bq. +  MapFSQueueType,SetString configuredQueues; Need a space after the 
comma.
Done

bq. Both NestedUserQueue and UserQueueBelow sound good.
Calling the rule 'NestedUserQueue'. Changed it at all places in the code.

I didn't add documentation in my last patch, I've changed FS doc in the 
latest(v4) patch.

 Fair Scheduler Dynamic Hierarchical User Queues
 ---

 Key: YARN-1864
 URL: https://issues.apache.org/jira/browse/YARN-1864
 Project: Hadoop YARN
  Issue Type: New Feature
  Components: scheduler
Reporter: Ashwin Shankar
  Labels: scheduler
 Attachments: YARN-1864-v1.txt, YARN-1864-v2.txt, YARN-1864-v3.txt, 
 YARN-1864-v4.txt


 In Fair Scheduler, we want to be able to create user queues under any parent 
 queue in the hierarchy. For eg. Say user1 submits a job to a parent queue 
 called root.allUserQueues, we want be able to create a new queue called 
 root.allUserQueues.user1 and run user1's job in it.Any further jobs submitted 
 by this user to root.allUserQueues will be run in this newly created 
 root.allUserQueues.user1.
 This is very similar to the 'user-as-default' feature in Fair Scheduler which 
 creates user queues under root queue. But we want the ability to create user 
 queues under ANY parent queue.
 Why do we want this ?
 1. Preemption : these dynamically created user queues can preempt each other 
 if its fair share is not met. So there is fairness among users.
 User queues can also preempt other non-user leaf queue as well if below fair 
 share.
 2. Allocation to user queues : we want all the user queries(adhoc) to consume 
 only a fraction of resources in the shared cluster. By creating this 
 feature,we could do that by giving a fair share to the parent user queue 
 which is then redistributed to all the dynamically created user queues.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987016#comment-13987016
 ] 

Vinod Kumar Vavilapalli commented on YARN-1201:
---

That doesn't seem to make sense. When would localhost be not resolvable?

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1945) Adding description for each pool in Fair Scheduler Page from fair-scheduler.xml

2014-05-01 Thread Siqi Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1945?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siqi Li updated YARN-1945:
--

Attachment: YARN-1945.v3.patch

 Adding description for each pool in Fair Scheduler Page from 
 fair-scheduler.xml
 ---

 Key: YARN-1945
 URL: https://issues.apache.org/jira/browse/YARN-1945
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 2.3.0
Reporter: Siqi Li
Assignee: Siqi Li
 Attachments: YARN-1945.v2.patch, YARN-1945.v3.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1354) Recover applications upon nodemanager restart

2014-05-01 Thread Jason Lowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1354?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe updated YARN-1354:
-

Attachment: YARN-1354-v2-and-YARN-1987-and-YARN-1362.patch

Updating the patch to address the DBException handling that was brought up in 
the MAPREDUCE-5652 review and applies here. Note that this now depends upon 
YARN-1987 as that provides the utility wrapper for the leveldb iterator to 
translate raw RuntimeException to the more helpful DBException so we can act 
accordingly when errors occur.

This patch also addresses the issue where apps were being cleaned up on 
shutdown.  This leverages YARN-1362 so we can distinguish a decommission 
shutdown, and it will avoid cleaning up applications if the state store can 
recover and we are not being decommissioned.  The other notable change in the 
patch is renaming LevelDB to Leveldb for consistency with the existing 
LeveldbTimelineStore naming convention.

This latest patch includes the necessary pieces of YARN-1987 and YARN-1362 so 
it can compile and Jenkins can comment.


 Recover applications upon nodemanager restart
 -

 Key: YARN-1354
 URL: https://issues.apache.org/jira/browse/YARN-1354
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: nodemanager
Affects Versions: 2.3.0
Reporter: Jason Lowe
Assignee: Jason Lowe
 Attachments: YARN-1354-v1.patch, 
 YARN-1354-v2-and-YARN-1987-and-YARN-1362.patch


 The set of active applications in the nodemanager context need to be 
 recovered for work-preserving nodemanager restart



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1987) Wrapper for leveldb DBIterator to aid in handling database exceptions

2014-05-01 Thread Ming Ma (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1987?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987059#comment-13987059
 ] 

Ming Ma commented on YARN-1987:
---

Jason, thanks for the clarification. The patch looks good to me.

 Wrapper for leveldb DBIterator to aid in handling database exceptions
 -

 Key: YARN-1987
 URL: https://issues.apache.org/jira/browse/YARN-1987
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 2.4.0
Reporter: Jason Lowe
Assignee: Jason Lowe
 Attachments: YARN-1987.patch


 Per discussions in YARN-1984 and MAPREDUCE-5652, it would be nice to have a 
 utility wrapper around leveldb's DBIterator to translate the raw 
 RuntimeExceptions it can throw into DBExceptions to make it easier to handle 
 database errors while iterating.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1713) Implement getnewapplication and submitapp as part of RM web service

2014-05-01 Thread shanyu zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987099#comment-13987099
 ] 

shanyu zhao commented on YARN-1713:
---

[~vvasudev] I tested this patch and it works great!

I thought about your implementation of reflecting how RPC structures the API 
and it sounds good to me. 

The only feedback I have is the style of the REST API, does the following API 
change make sense for you?
POST /app/id -- POST /appids
POST /app/{appid} -- POST /apps/{appid}
This way the submit app REST API is more consistent with app status and kill 
app API.

Sorry I missed the part that your implementation actually calls doAs() on 
behalf of the remote user. That should work just fine. 

Thanks!

 Implement getnewapplication and submitapp as part of RM web service
 ---

 Key: YARN-1713
 URL: https://issues.apache.org/jira/browse/YARN-1713
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Varun Vasudev
Assignee: Varun Vasudev
 Attachments: apache-yarn-1713.3.patch, apache-yarn-1713.4.patch, 
 apache-yarn-1713.5.patch, apache-yarn-1713.cumulative.2.patch, 
 apache-yarn-1713.cumulative.3.patch, apache-yarn-1713.cumulative.4.patch, 
 apache-yarn-1713.cumulative.patch, apache-yarn-1713.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Wangda Tan (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987102#comment-13987102
 ] 

Wangda Tan commented on YARN-1201:
--

[~vinodkv], localhost cannot be resolved would be happened when it doesn't 
contain a proper domain name, the mechanism used by Java to resolve a host is 
similar to nslookup hostname. Domain name not configured is a common case 
when user try Hadoop in local(vm).
As [~djp] suggested, I think we'd better throw a meaningful message rather than 
NPE when hostname cannot be resolved.

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Moved] (YARN-2014) Performance: AM scaleability is 10% slower in 2.4 compared to 0.23.9

2014-05-01 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2014?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli moved HADOOP-10553 to YARN-2014:


Affects Version/s: (was: 2.4.0)
   2.4.0
  Key: YARN-2014  (was: HADOOP-10553)
  Project: Hadoop YARN  (was: Hadoop Common)

 Performance: AM scaleability is 10% slower in 2.4 compared to 0.23.9
 

 Key: YARN-2014
 URL: https://issues.apache.org/jira/browse/YARN-2014
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: patrick white

 Performance comparison benchmarks from 2.x against 0.23 shows AM scalability 
 benchmark's runtime is approximately 10% slower in 2.4.0. The trend is 
 consistent across later releases in both lines, latest release numbers are:
 2.4.0.0 runtime 255.6 seconds (avg 5 passes)
 0.23.9.12 runtime 230.4 seconds (avg 5 passes)
 Diff: -9.9% 
 AM Scalability test is essentially a sleep job that measures time to launch 
 and complete a large number of mappers.
 The diff is consistent and has been reproduced in both a larger (350 node, 
 100,000 mappers) perf environment, as well as a small (10 node, 2,900 
 mappers) demo cluster.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-2014) Performance: AM scaleability is 10% slower in 2.4 compared to 0.23.9

2014-05-01 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2014?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-2014:
--

Target Version/s: 2.5.0

Tentatively setting target-version 2.5.0.

 Performance: AM scaleability is 10% slower in 2.4 compared to 0.23.9
 

 Key: YARN-2014
 URL: https://issues.apache.org/jira/browse/YARN-2014
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: patrick white

 Performance comparison benchmarks from 2.x against 0.23 shows AM scalability 
 benchmark's runtime is approximately 10% slower in 2.4.0. The trend is 
 consistent across later releases in both lines, latest release numbers are:
 2.4.0.0 runtime 255.6 seconds (avg 5 passes)
 0.23.9.12 runtime 230.4 seconds (avg 5 passes)
 Diff: -9.9% 
 AM Scalability test is essentially a sleep job that measures time to launch 
 and complete a large number of mappers.
 The diff is consistent and has been reproduced in both a larger (350 node, 
 100,000 mappers) perf environment, as well as a small (10 node, 2,900 
 mappers) demo cluster.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1368) Common work to re-populate containers’ state into scheduler

2014-05-01 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987155#comment-13987155
 ] 

Vinod Kumar Vavilapalli commented on YARN-1368:
---

bq. I think the prototype is doing as a scheduler-specifc fashion and only for 
the FairScheduler. It'll be a maintenance overhead if we implement each 
scheduler separately.
This makes sense. As much as possible, RM restart should be orthogonal to 
schedulers. Further, where possible, we should try to reuse code within the 
schedulers. That way we won't miss stuff like metrics, scheduler calculations, 
resource consumption etc.

 Common work to re-populate containers’ state into scheduler
 ---

 Key: YARN-1368
 URL: https://issues.apache.org/jira/browse/YARN-1368
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Bikas Saha
Assignee: Anubhav Dhoot
 Attachments: YARN-1368.preliminary.patch


 YARN-1367 adds support for the NM to tell the RM about all currently running 
 containers upon registration. The RM needs to send this information to the 
 schedulers along with the NODE_ADDED_EVENT so that the schedulers can recover 
 the current allocation state of the cluster.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1368) Common work to re-populate containers’ state into scheduler

2014-05-01 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987182#comment-13987182
 ] 

Sandy Ryza commented on YARN-1368:
--

+1 to sharing common code between schedulers where possible.  However, the 
patch proposed on this JIRA leaves two separate accountings for the nodes in 
the cluster - one inside AbstractYarnScheduler and one inside the specific 
schedulers.  As [~leftnoteasy] pointed out, if we want to merge scheduler code, 
it's probably best to do so in a patch that's separate from adding the 
re-population logic.

A meta-comment - [~jianhe], if you have concerns about an approach / prototype 
it's better to provide feedback than to rewrite it yourself.

 Common work to re-populate containers’ state into scheduler
 ---

 Key: YARN-1368
 URL: https://issues.apache.org/jira/browse/YARN-1368
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Bikas Saha
Assignee: Anubhav Dhoot
 Attachments: YARN-1368.preliminary.patch


 YARN-1367 adds support for the NM to tell the RM about all currently running 
 containers upon registration. The RM needs to send this information to the 
 schedulers along with the NODE_ADDED_EVENT so that the schedulers can recover 
 the current allocation state of the cluster.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1368) Common work to re-populate containers’ state into scheduler

2014-05-01 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1368?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987190#comment-13987190
 ] 

Vinod Kumar Vavilapalli commented on YARN-1368:
---

bq. Further, where possible, we should try to reuse code within the schedulers.
To clarify, what I meant was not reuse across schedulers (which BTW is in 
itself a good goal), but reusing existing code that is inside schedulers to 
also do stuff when recovering from a restart. Anyways..

 Common work to re-populate containers’ state into scheduler
 ---

 Key: YARN-1368
 URL: https://issues.apache.org/jira/browse/YARN-1368
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Bikas Saha
Assignee: Anubhav Dhoot
 Attachments: YARN-1368.preliminary.patch


 YARN-1367 adds support for the NM to tell the RM about all currently running 
 containers upon registration. The RM needs to send this information to the 
 schedulers along with the NODE_ADDED_EVENT so that the schedulers can recover 
 the current allocation state of the cluster.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-2012) Fair Scheduler : Default rule in queue placement policy can take a queue as an optional attribute

2014-05-01 Thread Ashwin Shankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2012?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ashwin Shankar updated YARN-2012:
-

Attachment: YARN-2012-v1.txt

 Fair Scheduler : Default rule in queue placement policy can take a queue as 
 an optional attribute
 -

 Key: YARN-2012
 URL: https://issues.apache.org/jira/browse/YARN-2012
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: scheduler
Reporter: Ashwin Shankar
  Labels: scheduler
 Attachments: YARN-2012-v1.txt


 Currently 'default' rule in queue placement policy,if applied,puts the app in 
 root.default queue. It would be great if we can make 'default' rule 
 optionally point to a different queue as default queue . This queue should be 
 an existing queue,if not we fall back to root.default queue hence keeping 
 this rule as terminal.
 This default queue can be a leaf queue or it can also be an parent queue if 
 the 'default' rule is nested inside nestedUserQueue rule(YARN-1864).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-2012) Fair Scheduler : Default rule in queue placement policy can take a queue as an optional attribute

2014-05-01 Thread Ashwin Shankar (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2012?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987205#comment-13987205
 ] 

Ashwin Shankar commented on YARN-2012:
--

Added an attribute 'queue' to default rule. Also added couple of unit tests.
Patch is expected get -1 from Hadoop QA,since this jira depends upon YARN-1864. 

 Fair Scheduler : Default rule in queue placement policy can take a queue as 
 an optional attribute
 -

 Key: YARN-2012
 URL: https://issues.apache.org/jira/browse/YARN-2012
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: scheduler
Reporter: Ashwin Shankar
  Labels: scheduler
 Attachments: YARN-2012-v1.txt


 Currently 'default' rule in queue placement policy,if applied,puts the app in 
 root.default queue. It would be great if we can make 'default' rule 
 optionally point to a different queue as default queue . This queue should be 
 an existing queue,if not we fall back to root.default queue hence keeping 
 this rule as terminal.
 This default queue can be a leaf queue or it can also be an parent queue if 
 the 'default' rule is nested inside nestedUserQueue rule(YARN-1864).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Wangda Tan (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987208#comment-13987208
 ] 

Wangda Tan commented on YARN-1201:
--

Uploaded a new patch to make this test case throw proper exception instead of a 
simple NPE. This test cases hard coded checking 
e.getCause().getMessage().contains(expectedMessage). When other exception 
thrown, like UnknownHostException, the e.getCause().getMessage is null. So 
updated type of exception cached from Exception to expected 
AccessControlException.

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-1201:
-

Attachment: YARN-1201.patch

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch, YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1961) Fair scheduler preemption doesn't work for non-leaf queues

2014-05-01 Thread Ashwin Shankar (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987221#comment-13987221
 ] 

Ashwin Shankar commented on YARN-1961:
--

hey [~sandyr], Do you know the history behind why preemption is implemented 
only at a leaf queue level ?
Isn't this a bug ?
Just wanted to get some context. 

 Fair scheduler preemption doesn't work for non-leaf queues
 --

 Key: YARN-1961
 URL: https://issues.apache.org/jira/browse/YARN-1961
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.4.0
Reporter: Ashwin Shankar
  Labels: scheduler

 Setting minResources and minSharePreemptionTimeout to a non-leaf queue 
 doesn't cause preemption to happen when that non-leaf queue is below 
 minResources and there are outstanding demands in that non-leaf queue.
 Here is an example fs allocation config(partial) :
 {code:xml}
 queue name=abc
   minResources3072 mb,0 vcores/minResources
   minSharePreemptionTimeout30/minSharePreemptionTimeout
 queue name=childabc1
 /queue
 queue name=childabc2
 /queue
  /queue
  {code}
 With the above configs,preemption doesn't seem to happen if queue abc is 
 below minShare and it has outstanding unsatisfied demands from apps in its 
 child queues. Ideally in such cases we would like preemption to kick off and 
 reclaim resources from other queues(not under queue abc).
 Looking at the code it seems like preemption checks for starvation only at 
 the leaf queue level and not at the parent level.
 {code:title=FairScheduler.java|borderStyle=solid}
 boolean isStarvedForMinShare(FSLeafQueue sched)
 boolean isStarvedForFairShare(FSLeafQueue sched)
 {code}
 This affects our use case where we have a parent queue with probably a 100 
 unconfigured leaf queues under it.We want to give a minshare to the parent 
 queue to protect all the leaf queues under it,but we cannot do it due to this 
 bug.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-2015) HTTPS doesn't work properly for daemons (RM, JHS, NM)

2014-05-01 Thread Robert Kanter (JIRA)
Robert Kanter created YARN-2015:
---

 Summary: HTTPS doesn't work properly for daemons (RM, JHS, NM)
 Key: YARN-2015
 URL: https://issues.apache.org/jira/browse/YARN-2015
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker


Enabling SSL in the site files and setting up a certificate, keystore, etc 
doesn't actually enable HTTPS.  The RM, NMs, and JHS will use their https port, 
but use only HTTP on them instead of only HTTPS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-2015) HTTPS doesn't work properly for daemons (RM, JHS, NM)

2014-05-01 Thread Robert Kanter (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2015?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987241#comment-13987241
 ] 

Robert Kanter commented on YARN-2015:
-

The problem appears to be that while HttpServer2 is able to run an HTTPS 
endpoint and everything (there's even a test for it), the configs are not 
hooked up to actually enable it.

 HTTPS doesn't work properly for daemons (RM, JHS, NM)
 -

 Key: YARN-2015
 URL: https://issues.apache.org/jira/browse/YARN-2015
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.4.0
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker

 Enabling SSL in the site files and setting up a certificate, keystore, etc 
 doesn't actually enable HTTPS.  The RM, NMs, and JHS will use their https 
 port, but use only HTTP on them instead of only HTTPS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-2015) HTTPS doesn't work properly for daemons (RM, JHS, NM)

2014-05-01 Thread Robert Kanter (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Kanter updated YARN-2015:


Affects Version/s: 2.3.0

 HTTPS doesn't work properly for daemons (RM, JHS, NM)
 -

 Key: YARN-2015
 URL: https://issues.apache.org/jira/browse/YARN-2015
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.3.0, 2.4.0
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker

 Enabling SSL in the site files and setting up a certificate, keystore, etc 
 doesn't actually enable HTTPS.  The RM, NMs, and JHS will use their https 
 port, but use only HTTP on them instead of only HTTPS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (YARN-2015) HTTPS doesn't work properly for daemons (RM, JHS, NM)

2014-05-01 Thread Robert Kanter (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Kanter resolved YARN-2015.
-

Resolution: Invalid

Nevermind, this appears to be fixed by YARN-1553

 HTTPS doesn't work properly for daemons (RM, JHS, NM)
 -

 Key: YARN-2015
 URL: https://issues.apache.org/jira/browse/YARN-2015
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.3.0, 2.4.0
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker

 Enabling SSL in the site files and setting up a certificate, keystore, etc 
 doesn't actually enable HTTPS.  The RM, NMs, and JHS will use their https 
 port, but use only HTTP on them instead of only HTTPS.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1339) Recover DeletionService state upon nodemanager restart

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1339?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987261#comment-13987261
 ] 

Hadoop QA commented on YARN-1339:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12642859/YARN-1339v3-and-YARN-1987.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 3 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3673//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3673//console

This message is automatically generated.

 Recover DeletionService state upon nodemanager restart
 --

 Key: YARN-1339
 URL: https://issues.apache.org/jira/browse/YARN-1339
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: nodemanager
Affects Versions: 2.3.0
Reporter: Jason Lowe
Assignee: Jason Lowe
 Attachments: YARN-1339.patch, YARN-1339v2.patch, 
 YARN-1339v3-and-YARN-1987.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987262#comment-13987262
 ] 

Hadoop QA commented on YARN-1201:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12642965/YARN-1201.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:

  
org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3671//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3671//console

This message is automatically generated.

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch, YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1945) Adding description for each pool in Fair Scheduler Page from fair-scheduler.xml

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1945?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987264#comment-13987264
 ] 

Hadoop QA commented on YARN-1945:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12642920/YARN-1945.v3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

  {color:red}-1 javac{color}.  The applied patch generated 1280 javac 
compiler warnings (more than the trunk's current 1278 warnings).

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3672//testReport/
Javac warnings: 
https://builds.apache.org/job/PreCommit-YARN-Build/3672//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3672//console

This message is automatically generated.

 Adding description for each pool in Fair Scheduler Page from 
 fair-scheduler.xml
 ---

 Key: YARN-1945
 URL: https://issues.apache.org/jira/browse/YARN-1945
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 2.3.0
Reporter: Siqi Li
Assignee: Siqi Li
 Attachments: YARN-1945.v2.patch, YARN-1945.v3.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-2012) Fair Scheduler : Default rule in queue placement policy can take a queue as an optional attribute

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2012?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987273#comment-13987273
 ] 

Hadoop QA commented on YARN-2012:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12642962/YARN-2012-v1.txt
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3676//console

This message is automatically generated.

 Fair Scheduler : Default rule in queue placement policy can take a queue as 
 an optional attribute
 -

 Key: YARN-2012
 URL: https://issues.apache.org/jira/browse/YARN-2012
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: scheduler
Reporter: Ashwin Shankar
  Labels: scheduler
 Attachments: YARN-2012-v1.txt


 Currently 'default' rule in queue placement policy,if applied,puts the app in 
 root.default queue. It would be great if we can make 'default' rule 
 optionally point to a different queue as default queue . This queue should be 
 an existing queue,if not we fall back to root.default queue hence keeping 
 this rule as terminal.
 This default queue can be a leaf queue or it can also be an parent queue if 
 the 'default' rule is nested inside nestedUserQueue rule(YARN-1864).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-1201:
-

Attachment: YARN-1201.patch

Fixed incorrect type import

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch, YARN-1201.patch, YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1995) YARN REST APIs section contains both YARN and MapReduce REST API pages

2014-05-01 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987329#comment-13987329
 ] 

Akira AJISAKA commented on YARN-1995:
-

Thanks for the report [~zjshen]. Since the issue is also reported by YARN-1999 
and there's more progress in the JIRA, would you please close this issue as 
duplicate?
I appreciate you will review the patch in YARN-1999.

 YARN REST APIs section contains both YARN and MapReduce REST API pages
 

 Key: YARN-1995
 URL: https://issues.apache.org/jira/browse/YARN-1995
 Project: Hadoop YARN
  Issue Type: Bug
  Components: site
Reporter: Zhijie Shen
Priority: Minor
  Labels: newbie, site

 To be clear to users, it's good to either put the MapReduce related pages in 
 a MAPREDUCE REST APIs section or rename current section name.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (YARN-1995) YARN REST APIs section contains both YARN and MapReduce REST API pages

2014-05-01 Thread Zhijie Shen (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijie Shen resolved YARN-1995.
---

Resolution: Duplicate

[~ajisakaa], thanks for informing me of the duplication. Will look at the patch.

 YARN REST APIs section contains both YARN and MapReduce REST API pages
 

 Key: YARN-1995
 URL: https://issues.apache.org/jira/browse/YARN-1995
 Project: Hadoop YARN
  Issue Type: Bug
  Components: site
Reporter: Zhijie Shen
Priority: Minor
  Labels: newbie, site

 To be clear to users, it's good to either put the MapReduce related pages in 
 a MAPREDUCE REST APIs section or rename current section name.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1201) TestAMAuthorization fails with local hostname cannot be resolved

2014-05-01 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1201?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13987357#comment-13987357
 ] 

Hadoop QA commented on YARN-1201:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12642986/YARN-1201.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:

  
org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3677//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3677//console

This message is automatically generated.

 TestAMAuthorization fails with local hostname cannot be resolved
 

 Key: YARN-1201
 URL: https://issues.apache.org/jira/browse/YARN-1201
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: SUSE Linux Enterprise Server 11 (x86_64)
Reporter: Nemon Lou
Assignee: Wangda Tan
Priority: Minor
 Attachments: YARN-1201.patch, YARN-1201.patch, YARN-1201.patch


 When hostname is 158-1-131-10, TestAMAuthorization fails.
 {code}
 Running org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 14.034 sec 
  FAILURE! - in 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization
 testUnauthorizedAccess[0](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.952 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 testUnauthorizedAccess[1](org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization)
   Time elapsed: 3.116 sec   ERROR!
 java.lang.NullPointerException: null
 at 
 org.apache.hadoop.yarn.server.resourcemanager.TestAMAuthorization.testUnauthorizedAccess(TestAMAuthorization.java:284)
 Results :
 Tests in error:
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
   TestAMAuthorization.testUnauthorizedAccess:284 NullPointer
 Tests run: 4, Failures: 0, Errors: 2, Skipped: 0
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)