[jira] [Updated] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Naganarasimha G R (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Naganarasimha G R updated YARN-3045:

Attachment: YARN-3045-YARN-2928.008.patch

updating the patch after  rebase with latest branch code ...

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3948) Display Application Priority in RM Web UI

2015-08-05 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3948?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14655153#comment-14655153
 ] 

Sunil G commented on YARN-3948:
---

FIndBugs report shown 0 warning.

> Display Application Priority in RM Web UI
> -
>
> Key: YARN-3948
> URL: https://issues.apache.org/jira/browse/YARN-3948
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: webapp
>Affects Versions: 2.7.1
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: 0001-YARN-3948.patch, 0002-YARN-3948.patch, 
> 0003-YARN-3948.patch, 0004-YARN-3948.patch, ApplicationPage.png, 
> ClusterPage.png
>
>
> Application Priority can be displayed in RM Web UI Application page.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-4019) Add JvmPauseMonitor to ResourceManager and NodeManager

2015-08-05 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14655239#comment-14655239
 ] 

Junping Du commented on YARN-4019:
--

Hi [~rkanter], thanks for contributing the patch! The Patch LGTM in overall. A 
small comment is: can we move below line from serviceInit() to serviceStart()?
{code}
+  pauseMonitor.start();
{code}
Other looks fine to me. I think the test failure is not related to your patch. 
Isn't it?

> Add JvmPauseMonitor to ResourceManager and NodeManager
> --
>
> Key: YARN-4019
> URL: https://issues.apache.org/jira/browse/YARN-4019
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: nodemanager, resourcemanager
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: YARN-4019.001.patch, YARN-4019.002.patch
>
>
> We should add the {{JvmPauseMonitor}} from HADOOP-9618 to the ResourceManager 
> and NodeManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658386#comment-14658386
 ] 

Junping Du commented on YARN-3045:
--

Thanks [~sjlee0] and [~Naganarasimha] for quickly reply.
bq. If these events are attributes of applications, then they should be on the 
application entities. If I want to find out all events for some application, 
then I should be able to query only the application entity and get all events.
Some of these events are related to both application and NodeManager. We can 
claim that it belongs to application but we can see that some events are too 
detailed to application but could be more interested for YARN daemons. I can 
understand that our design is more application centric now but should be 
generic enough to store/retrival YARN daemon centric entities later. Anyway, 
before making NM/RM onboard as the first class consumer of ATSv2, I am fine 
with making them as application events.

bq. The need to have NodeManagerEntity is something different IMO. Note that 
today there are challenges in emitting data without any application context 
(e.g. node manager's configuration) as we discussed a few times. If we need to 
support that, that needs a different discussion.
I see. I remember to see a JIRA work is to get ride of application context but 
cannot find it now. In case we don't have it, how about move this discussion to 
YARN-3959? The original scope of that JIRA is application related configuration 
only but we could extend it to include daemon configuration if necessary.

bq. my assumption was that the sync/async distinction from the client 
perspective mapped to whether the writer may be flushed or not. If not, then we 
need to support a 2x2 matrix of possibilities: sync put w/ flush, sync put w/o 
flush, async put w/ flush, and async put w/o flush. I thought it would be a 
simplifying assumption to align those dimensions.
I think we can simplify 2x2 matrix by omitting the case of sync put w/o flush 
as I cannot think a valid case that ack from TimelineCollector without flush 
can help on. Rest of three cases sounds solid to me. To make TimelineCollector 
can identify flush strategies with async calls, we may need to set severity on 
entities need to put and TimelineCollector is configured to flush entities only 
above specific severity just like log level does.

bq. I was under the impression that YARN-3367 is only for invoking REST calls 
in nonblocking way and thus avoiding threads in the clients. Is it also related 
to flush when called only putEntities and not on putEntitiesAsync?
You are right that the goal of YARN-3367 is to get rid of blocking call to put 
entities, no matter it calls putEntities() or something else. 
putEntitiesAsync() is exactly what we need, and it should be rare case to use 
putEntities() once we have putEntitiesAsync except client logic rely on return 
results tightly.

bq. I see currently "async" parameter as part of REST request is ignored now, 
so i thought based on this param we may need to further flush the writer or is 
your thoughts similar to support 2*2 matrix as Sangjin was informing?
Actually, from my above comments, I would prefer the way of (2*2 - 1). :) To 
speed up this JIRA's progress, I am fine with keep ignoring sync/async 
parameter and do everything async for now and left it out to a dedicated JIRA 
to figure out.

Will look at latest patch soon.

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3952) Fix new findbugs warnings in resourcemanager in YARN-2928 branch

2015-08-05 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658419#comment-14658419
 ] 

Junping Du commented on YARN-3952:
--

Great. Thx! Which jira patch you plan to have this fix? We can resolve this as 
duplicated to that one.

> Fix new findbugs warnings in resourcemanager in YARN-2928 branch
> 
>
> Key: YARN-3952
> URL: https://issues.apache.org/jira/browse/YARN-3952
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Affects Versions: YARN-2928
>Reporter: Varun Saxena
>Assignee: Varun Saxena
> Attachments: YARN-3952-YARN-2928.01.patch
>
>
> {noformat}
>  classname='org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher'>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='79'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptRegisteredEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='76'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationACLsUpdatedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='73'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='67'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationFinishedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='70'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='82'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='85'/>
> 
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3952) Fix new findbugs warnings in resourcemanager in YARN-2928 branch

2015-08-05 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3952?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658420#comment-14658420
 ] 

Junping Du commented on YARN-3952:
--

Great. Thx! Which jira patch you plan to have this fix? We can resolve this as 
duplicated to that one.

> Fix new findbugs warnings in resourcemanager in YARN-2928 branch
> 
>
> Key: YARN-3952
> URL: https://issues.apache.org/jira/browse/YARN-3952
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Affects Versions: YARN-2928
>Reporter: Varun Saxena
>Assignee: Varun Saxena
> Attachments: YARN-3952-YARN-2928.01.patch
>
>
> {noformat}
>  classname='org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher'>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='79'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptRegisteredEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='76'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationACLsUpdatedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='73'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='67'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationFinishedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='70'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='82'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='85'/>
> 
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Issue Comment Deleted] (YARN-3952) Fix new findbugs warnings in resourcemanager in YARN-2928 branch

2015-08-05 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3952?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated YARN-3952:
-
Comment: was deleted

(was: Great. Thx! Which jira patch you plan to have this fix? We can resolve 
this as duplicated to that one.)

> Fix new findbugs warnings in resourcemanager in YARN-2928 branch
> 
>
> Key: YARN-3952
> URL: https://issues.apache.org/jira/browse/YARN-3952
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Affects Versions: YARN-2928
>Reporter: Varun Saxena
>Assignee: Varun Saxena
> Attachments: YARN-3952-YARN-2928.01.patch
>
>
> {noformat}
>  classname='org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher'>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='79'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AppAttemptRegisteredEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='76'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationACLsUpdatedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='73'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='67'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ApplicationFinishedEvent
>  in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='70'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerCreatedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='82'/>
>  message='Unchecked/unconfirmed cast from 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.SystemMetricsEvent to 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.ContainerFinishedEvent 
> in 
> org.apache.hadoop.yarn.server.resourcemanager.metrics.AbstractTimelineServicePublisher.handle(SystemMetricsEvent)'
>  lineNumber='85'/>
> 
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3992) TestApplicationPriority.testApplicationPriorityAllocation fails intermittently

2015-08-05 Thread Sunil G (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3992?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sunil G updated YARN-3992:
--
Attachment: 0003-YARN-3992.patch

Thanks [~rohithsharma]. Change method signature so that existing test cases can 
use old method itself.

> TestApplicationPriority.testApplicationPriorityAllocation fails intermittently
> --
>
> Key: YARN-3992
> URL: https://issues.apache.org/jira/browse/YARN-3992
> Project: Hadoop YARN
>  Issue Type: Test
>Reporter: Zhijie Shen
>Assignee: Sunil G
> Attachments: 0001-YARN-3992.patch, 0002-YARN-3992.patch, 
> 0003-YARN-3992.patch
>
>
> {code}
> java.lang.AssertionError: expected:<7> but was:<5>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.TestApplicationPriority.testApplicationPriorityAllocation(TestApplicationPriority.java:182)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3992) TestApplicationPriority.testApplicationPriorityAllocation fails intermittently

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658535#comment-14658535
 ] 

Hadoop QA commented on YARN-3992:
-

\\
\\
| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |   7m  9s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 2 new or modified test files. |
| {color:green}+1{color} | javac |   8m 11s | There were no new javac warning 
messages. |
| {color:green}+1{color} | release audit |   0m 20s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 50s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 24s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   1m 32s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | yarn tests |  54m 20s | Tests passed in 
hadoop-yarn-server-resourcemanager. |
| | |  74m 29s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748870/0003-YARN-3992.patch |
| Optional Tests | javac unit findbugs checkstyle |
| git revision | trunk / 52f3525 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8772/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8772/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf905.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8772/console |


This message was automatically generated.

> TestApplicationPriority.testApplicationPriorityAllocation fails intermittently
> --
>
> Key: YARN-3992
> URL: https://issues.apache.org/jira/browse/YARN-3992
> Project: Hadoop YARN
>  Issue Type: Test
>Reporter: Zhijie Shen
>Assignee: Sunil G
> Attachments: 0001-YARN-3992.patch, 0002-YARN-3992.patch, 
> 0003-YARN-3992.patch
>
>
> {code}
> java.lang.AssertionError: expected:<7> but was:<5>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.TestApplicationPriority.testApplicationPriorityAllocation(TestApplicationPriority.java:182)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-2672) Improve Gridmix (synthetic generator + reservation support)

2015-08-05 Thread Carlo Curino (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2672?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Carlo Curino updated YARN-2672:
---
Attachment: YARN-2672.1.patch

Moving the patch forward to trunk "as is". It will need more "cleaning" before 
being committed, but people are using it / interested in it. So it is good to 
have it up to date.

> Improve Gridmix (synthetic generator + reservation support)
> ---
>
> Key: YARN-2672
> URL: https://issues.apache.org/jira/browse/YARN-2672
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Carlo Curino
>Assignee: Carlo Curino
>  Labels: BB2015-05-TBR
> Attachments: YARN-2672.1.patch, YARN-2672.patch
>
>
> This JIRA proposes an enhancement of Gridmix that contains:
> 1) a synthetic generator to produce load without the need of a trace, but 
> based on distributions
> 2) include negotiation of reservations (to test YARN-1051). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3487) CapacityScheduler scheduler lock obtained unnecessarily when calling getQueue

2015-08-05 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3487?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-3487:
-
Labels: 2.6.1-candidate  (was: )

> CapacityScheduler scheduler lock obtained unnecessarily when calling getQueue
> -
>
> Key: YARN-3487
> URL: https://issues.apache.org/jira/browse/YARN-3487
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Affects Versions: 2.6.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
>Priority: Critical
>  Labels: 2.6.1-candidate
> Fix For: 2.7.1
>
> Attachments: YARN-3487.001.patch, YARN-3487.002.patch, 
> YARN-3487.003.patch
>
>
> Recently saw a significant slowdown of applications on a large cluster, and 
> we noticed there were a large number of blocked threads on the RM.  Most of 
> the blocked threads were waiting for the CapacityScheduler lock while calling 
> getQueueInfo.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-2922) ConcurrentModificationException in CapacityScheduler's LeafQueue

2015-08-05 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-2922:
-
Labels: 2.6.1-candidate  (was: )

> ConcurrentModificationException in CapacityScheduler's LeafQueue
> 
>
> Key: YARN-2922
> URL: https://issues.apache.org/jira/browse/YARN-2922
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacityscheduler, resourcemanager, scheduler
>Affects Versions: 2.5.1
>Reporter: Jason Tufo
>Assignee: Rohith Sharma K S
>  Labels: 2.6.1-candidate
> Fix For: 2.7.0
>
> Attachments: 0001-YARN-2922.patch, 0001-YARN-2922.patch
>
>
> java.util.ConcurrentModificationException
> at 
> java.util.TreeMap$PrivateEntryIterator.nextEntry(TreeMap.java:1115)
> at java.util.TreeMap$KeyIterator.next(TreeMap.java:1169)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue.collectSchedulerApplications(LeafQueue.java:1618)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.getAppsInQueue(CapacityScheduler.java:1119)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.getQueueInfo(ClientRMService.java:798)
> at 
> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.getQueueInfo(ApplicationClientProtocolPBServiceImpl.java:234)
> at 
> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:333)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
> at 
> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
> at 
> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3487) CapacityScheduler scheduler lock obtained unnecessarily when calling getQueue

2015-08-05 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3487?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-3487:
-
Labels:   (was: 2.6.1-candidate)

> CapacityScheduler scheduler lock obtained unnecessarily when calling getQueue
> -
>
> Key: YARN-3487
> URL: https://issues.apache.org/jira/browse/YARN-3487
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Affects Versions: 2.6.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
>Priority: Critical
> Fix For: 2.7.1
>
> Attachments: YARN-3487.001.patch, YARN-3487.002.patch, 
> YARN-3487.003.patch
>
>
> Recently saw a significant slowdown of applications on a large cluster, and 
> we noticed there were a large number of blocked threads on the RM.  Most of 
> the blocked threads were waiting for the CapacityScheduler lock while calling 
> getQueueInfo.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-2922) ConcurrentModificationException in CapacityScheduler's LeafQueue

2015-08-05 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-2922:
-
Labels:   (was: 2.6.1-candidate)

> ConcurrentModificationException in CapacityScheduler's LeafQueue
> 
>
> Key: YARN-2922
> URL: https://issues.apache.org/jira/browse/YARN-2922
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacityscheduler, resourcemanager, scheduler
>Affects Versions: 2.5.1
>Reporter: Jason Tufo
>Assignee: Rohith Sharma K S
> Fix For: 2.7.0
>
> Attachments: 0001-YARN-2922.patch, 0001-YARN-2922.patch
>
>
> java.util.ConcurrentModificationException
> at 
> java.util.TreeMap$PrivateEntryIterator.nextEntry(TreeMap.java:1115)
> at java.util.TreeMap$KeyIterator.next(TreeMap.java:1169)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue.collectSchedulerApplications(LeafQueue.java:1618)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.getAppsInQueue(CapacityScheduler.java:1119)
> at 
> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.getQueueInfo(ClientRMService.java:798)
> at 
> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.getQueueInfo(ApplicationClientProtocolPBServiceImpl.java:234)
> at 
> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:333)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
> at 
> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
> at 
> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:415)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3988) DockerContainerExecutor should allow user specify "docker run" parameters

2015-08-05 Thread Chen He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3988?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658592#comment-14658592
 ] 

Chen He commented on YARN-3988:
---

I will post WIP patch, working on unit test, it requires docker ready on the 
compiling machine. 

> DockerContainerExecutor should allow user specify "docker run" parameters
> -
>
> Key: YARN-3988
> URL: https://issues.apache.org/jira/browse/YARN-3988
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.7.1
>Reporter: Chen He
>Assignee: Chen He
>
> In current DockerContainerExecutor, the "docker run" command has fixed 
> parameters:
> String commandStr = commands.append(dockerExecutor)
>   .append(" ")
>   .append("run")
>   .append(" ")
>   .append("--rm --net=host")
>   .append(" ")
>   .append(" --name " + containerIdStr)
>   .append(localDirMount)
>   .append(logDirMount)
>   .append(containerWorkDirMount)
>   .append(" ")
>   .append(containerImageName)
>   .toString();
> For example, it is not flexible if users want to start a docker container 
> with attaching extra volume(s) and other "docker run" parameters. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Anubhav Dhoot (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658619#comment-14658619
 ] 

Anubhav Dhoot commented on YARN-3736:
-

Findbugs result is clean 
https://builds.apache.org/job/PreCommit-YARN-Build/8765/artifact/patchprocess/patchFindbugsWarningshadoop-yarn-server-resourcemanager.html.

> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-2672) Improve Gridmix (synthetic generator + reservation support)

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658648#comment-14658648
 ] 

Hadoop QA commented on YARN-2672:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  16m 31s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 5 new or modified test files. |
| {color:green}+1{color} | javac |   8m 18s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 13s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 23s | The applied patch generated 
7 release audit warnings. |
| {color:red}-1{color} | checkstyle |   0m 31s | The applied patch generated  
177 new checkstyle issues (total was 126, now 300). |
| {color:red}-1{color} | whitespace |   0m  2s | The patch has 14  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 23s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 36s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   1m  1s | The patch appears to introduce 5 
new Findbugs (version 3.0.0) warnings. |
| {color:red}-1{color} | tools/hadoop tests |  11m 11s | Tests failed in 
hadoop-gridmix. |
| | |  50m 12s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-gridmix |
| Failed unit tests | hadoop.mapred.gridmix.TestGridmixMemoryEmulation |
|   | hadoop.mapred.gridmix.TestSleepJob |
|   | hadoop.mapred.gridmix.TestHighRamJob |
| Timed out tests | org.apache.hadoop.mapred.gridmix.TestGridmixSubmission |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748880/YARN-2672.1.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 4ab49a4 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-YARN-Build/8773/artifact/patchprocess/diffcheckstylehadoop-gridmix.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/artifact/patchprocess/whitespace.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/artifact/patchprocess/newPatchFindbugsWarningshadoop-gridmix.html
 |
| hadoop-gridmix test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/artifact/patchprocess/testrun_hadoop-gridmix.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf905.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8773/console |


This message was automatically generated.

> Improve Gridmix (synthetic generator + reservation support)
> ---
>
> Key: YARN-2672
> URL: https://issues.apache.org/jira/browse/YARN-2672
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Carlo Curino
>Assignee: Carlo Curino
>  Labels: BB2015-05-TBR
> Attachments: YARN-2672.1.patch, YARN-2672.patch
>
>
> This JIRA proposes an enhancement of Gridmix that contains:
> 1) a synthetic generator to produce load without the need of a trace, but 
> based on distributions
> 2) include negotiation of reservations (to test YARN-1051). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-4003) ReservationQueue inherit getAMResourceLimit() from LeafQueue, but behavior is not consistent

2015-08-05 Thread Carlo Curino (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4003?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658665#comment-14658665
 ] 

Carlo Curino commented on YARN-4003:


I think you are correct, but given that the size of the allocation for a 
ReservationQueue can fluctuate substantially, I think we need to rely on either 
hard-limits on the number of applications running per queue, or careful user 
submissions will need to protect from the bad scenario you describe. 
Picking the opposite strategy (using actual reservation size to limit number of 
AMs) prevents us to scavange resources early in the reservation, and I am not 
sure even works properly if the reservation "shrinks" (does ProportionalCPP 
kill AMs running in a queue, if the capacity is shrinked?).

I agree with your concerns, but I don't see a better way out.



> ReservationQueue inherit getAMResourceLimit() from LeafQueue, but behavior is 
> not consistent
> 
>
> Key: YARN-4003
> URL: https://issues.apache.org/jira/browse/YARN-4003
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Reporter: Carlo Curino
> Attachments: YARN-4003.patch
>
>
> The inherited behavior from LeafQueue (limit AM % based on capacity) is not a 
> good fit for ReservationQueue (that have highly dynamic capacity). 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Joep Rottinghuis (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658671#comment-14658671
 ] 

Joep Rottinghuis commented on YARN-3045:


Yeah, the discussion thread here seems to be rather deep. The first patch was 
put up in April and later versions of the patch seem to have been +1'ed already 
by several folks.
If there is a fundamental problem with the patch, we should address it.

It would be good to keep the comments here really focused on this patch and 
open separate jiras for separate topics to that we can keep making progress.

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3920) FairScheduler Reserving a node for a container should be configurable to allow it used only for large containers

2015-08-05 Thread Matthew Jacobs (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3920?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658674#comment-14658674
 ] 

Matthew Jacobs commented on YARN-3920:
--

[~adhoot]
{quote}
The problem is if you get that too high (such that it exceeds maximum resource 
allocation) one can accidentally disable reservation.
{quote}
That might be desired in some circumstances, no?

> FairScheduler Reserving a node for a container should be configurable to 
> allow it used only for large containers
> 
>
> Key: YARN-3920
> URL: https://issues.apache.org/jira/browse/YARN-3920
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: fairscheduler
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: yARN-3920.001.patch, yARN-3920.002.patch
>
>
> Reserving a node for a container was designed for preventing large containers 
> from starvation from small requests that keep getting into a node. Today we 
> let this be used even for a small container request. This has a huge impact 
> on scheduling since we block other scheduling requests until that reservation 
> is fulfilled. We should make this configurable so its impact can be minimized 
> by limiting it for large container requests as originally intended. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-4019) Add JvmPauseMonitor to ResourceManager and NodeManager

2015-08-05 Thread Robert Kanter (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-4019?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Kanter updated YARN-4019:

Attachment: YARN-4019.003.patch

Sure.  The 003 patch moves the start lines.  The test failure was unrelated.

> Add JvmPauseMonitor to ResourceManager and NodeManager
> --
>
> Key: YARN-4019
> URL: https://issues.apache.org/jira/browse/YARN-4019
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: nodemanager, resourcemanager
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: YARN-4019.001.patch, YARN-4019.002.patch, 
> YARN-4019.003.patch
>
>
> We should add the {{JvmPauseMonitor}} from HADOOP-9618 to the ResourceManager 
> and NodeManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658765#comment-14658765
 ] 

Arun Suresh commented on YARN-3736:
---

Looks good. Thanks for the patch [~adhoot] and the reviews [~subru]
Will be committing this shortly..

> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658773#comment-14658773
 ] 

Arun Suresh commented on YARN-3736:
---

Committed to trunk and branch-2

> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658798#comment-14658798
 ] 

Bikas Saha commented on YARN-3736:
--

Folks, how much load is this going to add to the state store (in terms of 
actual data volume and rate of stores/updates)? The original design of the RM 
state store assumes low volume of data and updates (proportional to number of 
new applications entering the system). So asking. Please let me know if this 
has been discussed and I missed that part.

> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3961) Expose queue container information (pending, running, reserved) in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun Suresh updated YARN-3961:
--
Attachment: YARN-3961.001.patch

Re-uploading last patch to kick jenkins

> Expose queue container information (pending, running, reserved) in REST api 
> and yarn top
> 
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658829#comment-14658829
 ] 

Hudson commented on YARN-3736:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8265 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8265/])
YARN-3736. Add RMStateStore apis to store and load accepted reservations for 
failover (adhoot via asuresh) (Arun Suresh: rev 
f271d377357ad680924d19f07e6c8315e7c89bae)
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/NullRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStoreEventType.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/ZKRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/proto/yarn_server_resourcemanager_recovery.proto
* hadoop-yarn-project/CHANGES.txt
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/reservation/ReservationSystemUtil.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/reservation/ReservationSystemTestUtil.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestLeveldbRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStoreStoreReservationEvent.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/reservation/TestInMemoryReservationAllocation.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStoreTestBase.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestFSRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/LeveldbRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/MemoryRMStateStore.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestZKRMStateStore.java


> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Ming Ma (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ming Ma updated YARN-221:
-
Attachment: YARN-221-8.patch

The javac warning isn't related to this patch. That is due to 
{{TestAuxServices}} cast Object to ArrayList. Updated the patch to 
take care of that anyway. The new patch also addresses the checkstyle and the 
whitespace issue.

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3983) Make CapacityScheduler to easier extend application allocation logic

2015-08-05 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658899#comment-14658899
 ] 

Jian He commented on YARN-3983:
---

failed test pass locally,  not related
committing

> Make CapacityScheduler to easier extend application allocation logic
> 
>
> Key: YARN-3983
> URL: https://issues.apache.org/jira/browse/YARN-3983
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Wangda Tan
>Assignee: Wangda Tan
> Fix For: 2.8.0
>
> Attachments: YARN-3983.1.patch, YARN-3983.2.patch, YARN-3983.3.patch, 
> YARN-3983.4.patch
>
>
> While working on YARN-1651 (resource allocation for increasing container), I 
> found it is very hard to extend existing CapacityScheduler resource 
> allocation logic to support different types of resource allocation.
> For example, there's a lot of differences between increasing a container and 
> allocating a container:
> - Increasing a container doesn't need to check locality delay.
> - Increasing a container doesn't need to build/modify a resource request tree 
> (ANY->RACK/HOST).
> - Increasing a container doesn't need to check allocation/reservation 
> starvation (see {{shouldAllocOrReserveNewContainer}}).
> - After increasing a container is approved by scheduler, it need to update an 
> existing container token instead of creating new container.
> And there're lots of similarities when allocating different types of 
> resources.
> - User-limit/queue-limit will be enforced for both of them.
> - Both of them needs resource reservation logic. (Maybe continuous 
> reservation looking is needed for both of them).
> The purpose of this JIRA is to make easier extending CapacityScheduler 
> resource allocation logic to support different types of resource allocation, 
> make common code reusable, and also better code organization.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-4019) Add JvmPauseMonitor to ResourceManager and NodeManager

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4019?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658943#comment-14658943
 ] 

Hadoop QA commented on YARN-4019:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  16m 25s | Findbugs (version ) appears to 
be broken on trunk. |
| {color:green}+1{color} | @author |   0m  1s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   8m  1s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 10s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 44s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 26s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 52s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:red}-1{color} | yarn tests |   5m 51s | Tests failed in 
hadoop-yarn-server-nodemanager. |
| {color:red}-1{color} | yarn tests |  51m 53s | Tests failed in 
hadoop-yarn-server-resourcemanager. |
| | |  98m 27s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.yarn.server.nodemanager.TestNodeManagerReboot |
|   | hadoop.yarn.server.nodemanager.TestDeletionService |
|   | hadoop.yarn.server.nodemanager.TestNodeManagerResync |
|   | 
hadoop.yarn.server.resourcemanager.scheduler.capacity.TestCapacitySchedulerQueueACLs
 |
|   | 
hadoop.yarn.server.resourcemanager.scheduler.capacity.TestNodeLabelContainerAllocation
 |
|   | hadoop.yarn.server.resourcemanager.TestRMAdminService |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748898/YARN-4019.003.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 4ab49a4 |
| hadoop-yarn-server-nodemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8774/artifact/patchprocess/testrun_hadoop-yarn-server-nodemanager.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8774/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8774/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf904.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8774/console |


This message was automatically generated.

> Add JvmPauseMonitor to ResourceManager and NodeManager
> --
>
> Key: YARN-4019
> URL: https://issues.apache.org/jira/browse/YARN-4019
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: nodemanager, resourcemanager
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: YARN-4019.001.patch, YARN-4019.002.patch, 
> YARN-4019.003.patch
>
>
> We should add the {{JvmPauseMonitor}} from HADOOP-9618 to the ResourceManager 
> and NodeManager.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3983) Make CapacityScheduler to easier extend application allocation logic

2015-08-05 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3983?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658955#comment-14658955
 ] 

Hudson commented on YARN-3983:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8266 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8266/])
YARN-3983. Refactored CapacityScheduleri#FiCaSchedulerApp to easier extend 
container allocation logic. Contributed by Wangda Tan (jianhe: rev 
ba2313d6145a1234777938a747187373f4cd58d9)
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/ParentQueue.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/allocator/AllocationState.java
* hadoop-yarn-project/CHANGES.txt
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/allocator/ContainerAllocator.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/LeafQueue.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestLeafQueue.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/allocator/RegularContainerAllocator.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/common/fica/FiCaSchedulerApp.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/allocator/ContainerAllocation.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CSAssignment.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/AbstractCSQueue.java


> Make CapacityScheduler to easier extend application allocation logic
> 
>
> Key: YARN-3983
> URL: https://issues.apache.org/jira/browse/YARN-3983
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Wangda Tan
>Assignee: Wangda Tan
> Fix For: 2.8.0
>
> Attachments: YARN-3983.1.patch, YARN-3983.2.patch, YARN-3983.3.patch, 
> YARN-3983.4.patch
>
>
> While working on YARN-1651 (resource allocation for increasing container), I 
> found it is very hard to extend existing CapacityScheduler resource 
> allocation logic to support different types of resource allocation.
> For example, there's a lot of differences between increasing a container and 
> allocating a container:
> - Increasing a container doesn't need to check locality delay.
> - Increasing a container doesn't need to build/modify a resource request tree 
> (ANY->RACK/HOST).
> - Increasing a container doesn't need to check allocation/reservation 
> starvation (see {{shouldAllocOrReserveNewContainer}}).
> - After increasing a container is approved by scheduler, it need to update an 
> existing container token instead of creating new container.
> And there're lots of similarities when allocating different types of 
> resources.
> - User-limit/queue-limit will be enforced for both of them.
> - Both of them needs resource reservation logic. (Maybe continuous 
> reservation looking is needed for both of them).
> The purpose of this JIRA is to make easier extending CapacityScheduler 
> resource allocation logic to support different types of resource allocation, 
> make common code reusable, and also better code organization.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14658963#comment-14658963
 ] 

Xuan Gong commented on YARN-221:


Thanks for the latest patch. I think that we are close. The patch looks good 
overall.  One nit:
*  could we modify this doc in AppLogAggregatorImpl, too
{code}
// Create a set of Containers whose logs will be uploaded in this cycle.
// It includes:
// a) all containers in pendingContainers: those containers are finished
//and satisfy the retentionPolicy.
// b) some set of running containers: For all the Running containers,
// we have ContainerLogsRetentionPolicy.AM_AND_FAILED_CONTAINERS_ONLY,
// so simply set wasContainerSuccessful as true to
// bypass FAILED_CONTAINERS check and find the running containers 
// which satisfy the retentionPolicy.
{code}

Also, I realized that ContainerTokenIdentifier is used here
{code}
boolean shouldDoLogAggregation(ContainerTokenIdentifier containerToken,  int 
exitCode);
{code}
Currently, it is fine. But if in future, we might need other information which 
the ContainerTokenIdentifier can not provide. So, probably, we could have our 
own ContainerLogContext instead of using ContainerTokenIdentifier ? In that 
case, if we have requirement to use other information, we could add. 

Thoughts ?

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-221:
---
Labels: new  (was: )

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
>  Labels: new
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-221:
---
Labels:   (was: new)

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3049) [Storage Implementation] Implement storage reader interface to fetch raw data from HBase backend

2015-08-05 Thread Li Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659012#comment-14659012
 ] 

Li Lu commented on YARN-3049:
-

Hi [~zjshen], letting HBase implementation locally looks good to me. One minor 
comment for the latest patch is, maybe we want to separate the logic like {{if 
(te.getType().equals(TimelineEntityType.YARN_APPLICATION.toString()))}} in 
HBaseWriterImpl into a separate private method? I think it will be much clearer 
to say something like:
{code}
if (te.getType().equals(TimelineEntityType.YARN_APPLICATION.toString())) {
  updateAppToFlowTable(te);
}
{code}

As [~sjlee0] mentioned above that we may have some other specialization within 
HBaseWriterImpl, so maybe it's helpful to let these special designs stand out? 

> [Storage Implementation] Implement storage reader interface to fetch raw data 
> from HBase backend
> 
>
> Key: YARN-3049
> URL: https://issues.apache.org/jira/browse/YARN-3049
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Zhijie Shen
> Attachments: YARN-3049-WIP.1.patch, YARN-3049-WIP.2.patch, 
> YARN-3049-WIP.3.patch, YARN-3049-YARN-2928.2.patch, 
> YARN-3049-YARN-2928.3.patch, YARN-3049-YARN-2928.4.patch, 
> YARN-3049-YARN-2928.5.patch, YARN-3049-YARN-2928.6.patch
>
>
> Implement existing ATS queries with the new ATS reader design.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3904) Refactor timelineservice.storage to add support to online and offline aggregation writers

2015-08-05 Thread Li Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3904?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659016#comment-14659016
 ] 

Li Lu commented on YARN-3904:
-

Thanks [~sjlee0]! Any other comments from anyone? This JIRA is currently 
blocking the POC patch of YARN-3817. 

> Refactor timelineservice.storage to add support to online and offline 
> aggregation writers
> -
>
> Key: YARN-3904
> URL: https://issues.apache.org/jira/browse/YARN-3904
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Li Lu
>Assignee: Li Lu
> Attachments: YARN-3904-YARN-2928.001.patch, 
> YARN-3904-YARN-2928.002.patch, YARN-3904-YARN-2928.003.patch, 
> YARN-3904-YARN-2928.004.patch, YARN-3904-YARN-2928.005.patch, 
> YARN-3904-YARN-2928.006.patch, YARN-3904-YARN-2928.007.patch
>
>
> After we finished the design for time-based aggregation, we can adopt our 
> existing Phoenix storage into the storage of the aggregated data. In this 
> JIRA, I'm proposing to refactor writers to add support to aggregation 
> writers. Offline aggregation writers typically has less contextual 
> information. We can distinguish these writers by special naming. We can also 
> use CollectorContexts to model all contextual information and use it in our 
> writer interfaces. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3920) FairScheduler Reserving a node for a container should be configurable to allow it used only for large containers

2015-08-05 Thread Anubhav Dhoot (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3920?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659048#comment-14659048
 ] 

Anubhav Dhoot commented on YARN-3920:
-

Yeah that should be possible to do when thats the intention and thats ok. I am 
trying to avoid cases when its not the intention to disable reservation but it 
happens due to environment+configuration. 
In ratio of max case, one can disable reservation only by setting the ratio to 
above 1.0. 
In the multiple of increment case, one can disable it by setting say setting a 
multiple to  say 4 times increment and your max on the cluster (which depends 
on the min node size of cluster in addition to max config) is 3 times the 
increment. That can happen even if it was not your intention to disable it.


> FairScheduler Reserving a node for a container should be configurable to 
> allow it used only for large containers
> 
>
> Key: YARN-3920
> URL: https://issues.apache.org/jira/browse/YARN-3920
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: fairscheduler
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: yARN-3920.001.patch, yARN-3920.002.patch
>
>
> Reserving a node for a container was designed for preventing large containers 
> from starvation from small requests that keep getting into a node. Today we 
> let this be used even for a small container request. This has a huge impact 
> on scheduling since we block other scheduling requests until that reservation 
> is fulfilled. We should make this configurable so its impact can be minimized 
> by limiting it for large container requests as originally intended. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3920) FairScheduler Reserving a node for a container should be configurable to allow it used only for large containers

2015-08-05 Thread Anubhav Dhoot (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anubhav Dhoot updated YARN-3920:

Attachment: yARN-3920.003.patch

Attaching patch that updates reservation threshold in sync with changes in the 
max resource

> FairScheduler Reserving a node for a container should be configurable to 
> allow it used only for large containers
> 
>
> Key: YARN-3920
> URL: https://issues.apache.org/jira/browse/YARN-3920
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: fairscheduler
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: yARN-3920.001.patch, yARN-3920.002.patch, 
> yARN-3920.003.patch
>
>
> Reserving a node for a container was designed for preventing large containers 
> from starvation from small requests that keep getting into a node. Today we 
> let this be used even for a small container request. This has a huge impact 
> on scheduling since we block other scheduling requests until that reservation 
> is fulfilled. We should make this configurable so its impact can be minimized 
> by limiting it for large container requests as originally intended. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3736) Add RMStateStore apis to store and load accepted reservations for failover

2015-08-05 Thread Subru Krishnan (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3736?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659080#comment-14659080
 ] 

Subru Krishnan commented on YARN-3736:
--

Thanks [~adhoot] for the patch and [~asuresh] for reviewing/committing the 
patch.

[~bikassaha], we did have this discussion. The number of new reservations will 
be bounded by the number of applications as having a reservation per 
application is the worst case scenario. Practically you will have a reservation 
per job pipeline as a production pipeline is rarely composed of one job. 
Similarly Hive/Pig/Tez DAGs which are composed of multiple YARN apps can also 
be represented as a single reservation. Moreover not all new applications will 
use reservations, just those which need SLA and those generally tend to be 
bigger but fewer. Considering all this, our approach is aligned with the design 
of the state store. Makes sense?

> Add RMStateStore apis to store and load accepted reservations for failover
> --
>
> Key: YARN-3736
> URL: https://issues.apache.org/jira/browse/YARN-3736
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: YARN-3736.001.patch, YARN-3736.001.patch, 
> YARN-3736.002.patch, YARN-3736.003.patch, YARN-3736.004.patch, 
> YARN-3736.005.patch
>
>
> We need to persist the current state of the plan, i.e. the accepted 
> ReservationAllocations & corresponding RLESpareseResourceAllocations  to the 
> RMStateStore so that we can recover them on RM failover. This involves making 
> all the reservation system data structures protobuf friendly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3961) Expose queue container information (pending, running, reserved) in REST api and yarn top

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659109#comment-14659109
 ] 

Hadoop QA commented on YARN-3961:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  19m 51s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 43s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 42s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   2m 40s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  1s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 23s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   5m 23s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | yarn tests |   0m 23s | Tests passed in 
hadoop-yarn-api. |
| {color:green}+1{color} | yarn tests |   6m 57s | Tests passed in 
hadoop-yarn-client. |
| {color:green}+1{color} | yarn tests |   2m 10s | Tests passed in 
hadoop-yarn-common. |
| {color:red}-1{color} | yarn tests |  54m 49s | Tests failed in 
hadoop-yarn-server-resourcemanager. |
| | | 112m  4s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | 
hadoop.yarn.server.resourcemanager.webapp.TestRMWebServicesCapacitySched |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748913/YARN-3961.001.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / f271d37 |
| hadoop-yarn-api test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/artifact/patchprocess/testrun_hadoop-yarn-api.txt
 |
| hadoop-yarn-client test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/artifact/patchprocess/testrun_hadoop-yarn-client.txt
 |
| hadoop-yarn-common test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/artifact/patchprocess/testrun_hadoop-yarn-common.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf903.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8775/console |


This message was automatically generated.

> Expose queue container information (pending, running, reserved) in REST api 
> and yarn top
> 
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Li Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659121#comment-14659121
 ] 

Li Lu commented on YARN-3045:
-

bq. To speed up this JIRA's progress, I am fine with keep ignoring sync/async 
parameter and do everything async for now and left it out to a dedicated JIRA 
to figure out.

+1. This JIRA has been hanging there for quite a while. Let's move forward with 
pending storage API problems addressed in separate JIRAs. 

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3984) Rethink event column key issue

2015-08-05 Thread Zhijie Shen (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3984?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659183#comment-14659183
 ] 

Zhijie Shen commented on YARN-3984:
---

bq. If the info map is not empty, this record would be redundant and will take 
up storage space.

Make sense. The patch looks good to me. Will commit it.

> Rethink event column key issue
> --
>
> Key: YARN-3984
> URL: https://issues.apache.org/jira/browse/YARN-3984
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Zhijie Shen
>Assignee: Vrushali C
> Fix For: YARN-2928
>
> Attachments: YARN-3984-YARN-2928.001.patch
>
>
> Currently, the event column key is event_id?info_key?timestamp, which is not 
> so friendly to fetching all the events of an entity and sorting them in a 
> chronologic order. IMHO, timestamp?event_id?info_key may be a better key 
> schema. I open this jira to continue the discussion about it which was 
> commented on YARN-3908.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (YARN-4020) Exception happens while stopContainer in AM

2015-08-05 Thread sandflee (JIRA)
sandflee created YARN-4020:
--

 Summary: Exception happens while stopContainer in AM
 Key: YARN-4020
 URL: https://issues.apache.org/jira/browse/YARN-4020
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: sandflee


work-preserving-enable, RM HA, NM restart all enabled. The AM has running for a 
few weeks and when using NMClient.stopContainer, the following exception 
happens:

WARN ipc.Client: Exception encountered while connecting to the server : 
org.apache.hadoop.ipc.RemoteException(javax.security.sasl.SaslException): 
DIGEST-MD5: digest response format violation. Mismatched response.




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659204#comment-14659204
 ] 

Hadoop QA commented on YARN-221:


\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  21m 47s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 4 new or modified test files. |
| {color:green}+1{color} | javac |   7m 42s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 41s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   3m 21s | The applied patch generated  1 
new checkstyle issues (total was 212, now 212). |
| {color:green}+1{color} | whitespace |   1m 49s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 22s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   7m 39s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | common tests |  22m 29s | Tests passed in 
hadoop-common. |
| {color:green}+1{color} | yarn tests |   0m 23s | Tests passed in 
hadoop-yarn-api. |
| {color:green}+1{color} | yarn tests |   1m 55s | Tests passed in 
hadoop-yarn-common. |
| {color:green}+1{color} | yarn tests |   7m 35s | Tests passed in 
hadoop-yarn-server-nodemanager. |
| {color:green}+1{color} | yarn tests |  52m 48s | Tests passed in 
hadoop-yarn-server-resourcemanager. |
| | | 140m 16s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748921/YARN-221-8.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / ba2313d |
| checkstyle |  
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/diffcheckstylehadoop-yarn-api.txt
 |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/testrun_hadoop-common.txt
 |
| hadoop-yarn-api test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/testrun_hadoop-yarn-api.txt
 |
| hadoop-yarn-common test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/testrun_hadoop-yarn-common.txt
 |
| hadoop-yarn-server-nodemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/testrun_hadoop-yarn-server-nodemanager.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8776/console |


This message was automatically generated.

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (YARN-4021) RuntimeException/YarnRuntimeException sent over to the client can cause client to assume a local fatal failure

2015-08-05 Thread Anubhav Dhoot (JIRA)
Anubhav Dhoot created YARN-4021:
---

 Summary: RuntimeException/YarnRuntimeException sent over to the 
client can cause client to assume a local fatal failure 
 Key: YARN-4021
 URL: https://issues.apache.org/jira/browse/YARN-4021
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Anubhav Dhoot
Assignee: Anubhav Dhoot


Currently RuntimeException and its derived types such as YarnRuntimeExceptions 
are serialized over to the client and thrown at the client after YARN-731. This 
can cause issues like MAPREDUCE-6439 where we assume a local fatal exception 
has happened. 
Instead we should have a way to distinguish local RuntimeException versus 
remote RuntimeException to avoid these issues. We need to go over all the 
current client side code that is expecting a remote RuntimeException inorder to 
make it work with this change.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3961) Expose queue container information (pending, running, reserved) in REST api and yarn top

2015-08-05 Thread Anubhav Dhoot (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anubhav Dhoot updated YARN-3961:

Attachment: YARN-3961.002.patch

Fixed the test case failure

> Expose queue container information (pending, running, reserved) in REST api 
> and yarn top
> 
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3049) [Storage Implementation] Implement storage reader interface to fetch raw data from HBase backend

2015-08-05 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659256#comment-14659256
 ] 

Sangjin Lee commented on YARN-3049:
---

The latest patch looks good to me overall. Just a couple of comments.

I concur with [~gtCarrera9] that it might be a good idea to create more 
abstract methods around it. Note that we may be writing to other tables at this 
point too. We can even create private helper methods that check whether the 
entity is an application and so on. It's not critical but could be helpful...

Also, in {{HBaseTimelineWriterImpl}}, I see that the app-to-flow table is not 
being flushed. Either we should flush at the end of the write, or add it to the 
{{flush()}} method.

> [Storage Implementation] Implement storage reader interface to fetch raw data 
> from HBase backend
> 
>
> Key: YARN-3049
> URL: https://issues.apache.org/jira/browse/YARN-3049
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Zhijie Shen
> Attachments: YARN-3049-WIP.1.patch, YARN-3049-WIP.2.patch, 
> YARN-3049-WIP.3.patch, YARN-3049-YARN-2928.2.patch, 
> YARN-3049-YARN-2928.3.patch, YARN-3049-YARN-2928.4.patch, 
> YARN-3049-YARN-2928.5.patch, YARN-3049-YARN-2928.6.patch
>
>
> Implement existing ATS queries with the new ATS reader design.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659266#comment-14659266
 ] 

Sangjin Lee commented on YARN-3045:
---

bq. I see. I remember to see a JIRA work is to get ride of application context 
but cannot find it now.

It's YARN-3981.

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3974) Refactor the reservation system test cases to use parameterized base test

2015-08-05 Thread Subru Krishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3974?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Subru Krishnan updated YARN-3974:
-
Attachment: YARN-3974-v5.patch

Rebasing with trunk

> Refactor the reservation system test cases to use parameterized base test
> -
>
> Key: YARN-3974
> URL: https://issues.apache.org/jira/browse/YARN-3974
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler
>Reporter: Subru Krishnan
>Assignee: Subru Krishnan
> Attachments: YARN-3974-v1.patch, YARN-3974-v2.patch, 
> YARN-3974-v3.patch, YARN-3974-v4.patch, YARN-3974-v5.patch
>
>
> We have two test suites for testing ReservationSystem against Capacity & Fair 
> scheduler. We should combine them using a parametrized reservation system 
> base test similar to YARN-2797



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-221) NM should provide a way for AM to tell it not to aggregate logs.

2015-08-05 Thread Ming Ma (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659286#comment-14659286
 ] 

Ming Ma commented on YARN-221:
--

That sounds a good idea. How about using the existing 
ContainerTerminationContext? We can extend that to include exitCode. In that 
way, we don't need to introduce another somewhat similar Context class.

> NM should provide a way for AM to tell it not to aggregate logs.
> 
>
> Key: YARN-221
> URL: https://issues.apache.org/jira/browse/YARN-221
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: log-aggregation, nodemanager
>Reporter: Robert Joseph Evans
>Assignee: Ming Ma
> Attachments: YARN-221-6.patch, YARN-221-7.patch, YARN-221-8.patch, 
> YARN-221-trunk-v1.patch, YARN-221-trunk-v2.patch, YARN-221-trunk-v3.patch, 
> YARN-221-trunk-v4.patch, YARN-221-trunk-v5.patch
>
>
> The NodeManager should provide a way for an AM to tell it that either the 
> logs should not be aggregated, that they should be aggregated with a high 
> priority, or that they should be aggregated but with a lower priority.  The 
> AM should be able to do this in the ContainerLaunch context to provide a 
> default value, but should also be able to update the value when the container 
> is released.
> This would allow for the NM to not aggregate logs in some cases, and avoid 
> connection to the NN at all.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Naganarasimha G R (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659301#comment-14659301
 ] 

Naganarasimha G R commented on YARN-3045:
-

[~djp] & [~sjlee0]
Waiting for comments on the approach taken and wanted what exact aspects of 
localization needs to be captured. Is it related to 
{{ResourceLocalizationService events, i.e. 
LocalizationEventType.INIT_CONTAINER_RESOURCES & 
CONTAINER_RESOURCES_LOCALIZED}} / {{events of each individual Localized 
Resource i.e. ResourceEventType.REQUEST, LOCALIZED & LOCALIZATION_FAILED}} / 
{{ContainerEventType.RESOURCE_LOCALIZED &  RESOURCE_FAILED}} is sufficient to 
be captured ? 
In my opinion last option is sufficient or shall we handle localization events 
in another jira. Please share your thoughts ?

bq. . Anyway, before making NM/RM onboard as the first class consumer of ATSv2, 
I am fine with making them as application events.
Ok, going ahead to capture them under Application Entity.


> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3045) [Event producers] Implement NM writing container lifecycle events to ATS

2015-08-05 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659323#comment-14659323
 ] 

Sangjin Lee commented on YARN-3045:
---

bq. In my opinion last option is sufficient or shall we handle localization 
events in another jira. Please share your thoughts ?

Sorry but it's not clear what the 2 options are. Could you kindly rephrase the 
options?

Also, one clarifying question. Are some of these events already existing 
container events? If so, they shouldn't be repeated as application events 
redundantly, right? What would be the application-specific events that are 
*not* captured by container events?

> [Event producers] Implement NM writing container lifecycle events to ATS
> 
>
> Key: YARN-3045
> URL: https://issues.apache.org/jira/browse/YARN-3045
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Naganarasimha G R
> Attachments: YARN-3045-YARN-2928.002.patch, 
> YARN-3045-YARN-2928.003.patch, YARN-3045-YARN-2928.004.patch, 
> YARN-3045-YARN-2928.005.patch, YARN-3045-YARN-2928.006.patch, 
> YARN-3045-YARN-2928.007.patch, YARN-3045-YARN-2928.008.patch, 
> YARN-3045.20150420-1.patch
>
>
> Per design in YARN-2928, implement NM writing container lifecycle events and 
> container system metrics to ATS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3049) [Storage Implementation] Implement storage reader interface to fetch raw data from HBase backend

2015-08-05 Thread Zhijie Shen (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijie Shen updated YARN-3049:
--
Attachment: YARN-3049-YARN-2928.7.patch

Attach a new patch:

1. Rebase against YARN-3984.
2. Address Sangjin and Li's comments.

There's still a remaining issue: the timestamp will not be ser/des correctly by 
using UTF-8. I didn't figure the reason, but I did an experiment that the bytes 
were converted into string and then bytes, and they became different. Still 
need to do more investigation about this problem.

> [Storage Implementation] Implement storage reader interface to fetch raw data 
> from HBase backend
> 
>
> Key: YARN-3049
> URL: https://issues.apache.org/jira/browse/YARN-3049
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Zhijie Shen
> Attachments: YARN-3049-WIP.1.patch, YARN-3049-WIP.2.patch, 
> YARN-3049-WIP.3.patch, YARN-3049-YARN-2928.2.patch, 
> YARN-3049-YARN-2928.3.patch, YARN-3049-YARN-2928.4.patch, 
> YARN-3049-YARN-2928.5.patch, YARN-3049-YARN-2928.6.patch, 
> YARN-3049-YARN-2928.7.patch
>
>
> Implement existing ATS queries with the new ATS reader design.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3961) Expose queue container information (pending, running, reserved) in REST api and yarn top

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659383#comment-14659383
 ] 

Hadoop QA commented on YARN-3961:
-

\\
\\
| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  21m  1s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   8m 15s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 11s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   2m 56s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  1s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 29s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 36s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   5m 47s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | yarn tests |   0m 25s | Tests passed in 
hadoop-yarn-api. |
| {color:green}+1{color} | yarn tests |   6m 58s | Tests passed in 
hadoop-yarn-client. |
| {color:green}+1{color} | yarn tests |   2m  1s | Tests passed in 
hadoop-yarn-common. |
| {color:green}+1{color} | yarn tests |  53m 26s | Tests passed in 
hadoop-yarn-server-resourcemanager. |
| | | 113m 33s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748960/YARN-3961.002.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / f59612e |
| hadoop-yarn-api test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/artifact/patchprocess/testrun_hadoop-yarn-api.txt
 |
| hadoop-yarn-client test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/artifact/patchprocess/testrun_hadoop-yarn-client.txt
 |
| hadoop-yarn-common test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/artifact/patchprocess/testrun_hadoop-yarn-common.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf905.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8777/console |


This message was automatically generated.

> Expose queue container information (pending, running, reserved) in REST api 
> and yarn top
> 
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3974) Refactor the reservation system test cases to use parameterized base test

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3974?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659409#comment-14659409
 ] 

Hadoop QA commented on YARN-3974:
-

\\
\\
| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  16m 14s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 6 new or modified test files. |
| {color:green}+1{color} | javac |   7m 46s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 40s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 56s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  2s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 21s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 36s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   1m 27s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | yarn tests |  52m 38s | Tests passed in 
hadoop-yarn-server-resourcemanager. |
| | |  91m  6s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748970/YARN-3974-v5.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / f59612e |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8778/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8778/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8778/console |


This message was automatically generated.

> Refactor the reservation system test cases to use parameterized base test
> -
>
> Key: YARN-3974
> URL: https://issues.apache.org/jira/browse/YARN-3974
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler, fairscheduler
>Reporter: Subru Krishnan
>Assignee: Subru Krishnan
> Attachments: YARN-3974-v1.patch, YARN-3974-v2.patch, 
> YARN-3974-v3.patch, YARN-3974-v4.patch, YARN-3974-v5.patch
>
>
> We have two test suites for testing ReservationSystem against Capacity & Fair 
> scheduler. We should combine them using a parametrized reservation system 
> base test similar to YARN-2797



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3049) [Storage Implementation] Implement storage reader interface to fetch raw data from HBase backend

2015-08-05 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659413#comment-14659413
 ] 

Hadoop QA commented on YARN-3049:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 11s | Findbugs (version ) appears to 
be broken on YARN-2928. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 48s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 49s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   1m 17s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m 11s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 29s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 40s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 20s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | yarn tests |   0m 24s | Tests passed in 
hadoop-yarn-api. |
| {color:green}+1{color} | yarn tests |   1m 24s | Tests passed in 
hadoop-yarn-server-timelineservice. |
| | |  43m  2s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12748985/YARN-3049-YARN-2928.7.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | YARN-2928 / 895ccfa |
| hadoop-yarn-api test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8779/artifact/patchprocess/testrun_hadoop-yarn-api.txt
 |
| hadoop-yarn-server-timelineservice test log | 
https://builds.apache.org/job/PreCommit-YARN-Build/8779/artifact/patchprocess/testrun_hadoop-yarn-server-timelineservice.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/8779/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf906.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/8779/console |


This message was automatically generated.

> [Storage Implementation] Implement storage reader interface to fetch raw data 
> from HBase backend
> 
>
> Key: YARN-3049
> URL: https://issues.apache.org/jira/browse/YARN-3049
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Sangjin Lee
>Assignee: Zhijie Shen
> Attachments: YARN-3049-WIP.1.patch, YARN-3049-WIP.2.patch, 
> YARN-3049-WIP.3.patch, YARN-3049-YARN-2928.2.patch, 
> YARN-3049-YARN-2928.3.patch, YARN-3049-YARN-2928.4.patch, 
> YARN-3049-YARN-2928.5.patch, YARN-3049-YARN-2928.6.patch, 
> YARN-3049-YARN-2928.7.patch
>
>
> Implement existing ATS queries with the new ATS reader design.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3992) TestApplicationPriority.testApplicationPriorityAllocation fails intermittently

2015-08-05 Thread Rohith Sharma K S (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659491#comment-14659491
 ] 

Rohith Sharma K S commented on YARN-3992:
-

+1 lgtm

> TestApplicationPriority.testApplicationPriorityAllocation fails intermittently
> --
>
> Key: YARN-3992
> URL: https://issues.apache.org/jira/browse/YARN-3992
> Project: Hadoop YARN
>  Issue Type: Test
>Reporter: Zhijie Shen
>Assignee: Sunil G
> Attachments: 0001-YARN-3992.patch, 0002-YARN-3992.patch, 
> 0003-YARN-3992.patch
>
>
> {code}
> java.lang.AssertionError: expected:<7> but was:<5>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.TestApplicationPriority.testApplicationPriorityAllocation(TestApplicationPriority.java:182)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3992) TestApplicationPriority.testApplicationPriorityAllocation fails intermittently

2015-08-05 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659514#comment-14659514
 ] 

Hudson commented on YARN-3992:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8269 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8269/])
YARN-3992. TestApplicationPriority.testApplicationPriorityAllocation fails 
intermittently. (Contributed by Sunil G) (rohithsharmaks: rev 
df9e7280db58baddd02d6e23d3685efb8d5f1b97)
* hadoop-yarn-project/CHANGES.txt
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestApplicationPriority.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockAM.java


> TestApplicationPriority.testApplicationPriorityAllocation fails intermittently
> --
>
> Key: YARN-3992
> URL: https://issues.apache.org/jira/browse/YARN-3992
> Project: Hadoop YARN
>  Issue Type: Test
>Reporter: Zhijie Shen
>Assignee: Sunil G
> Fix For: 2.8.0
>
> Attachments: 0001-YARN-3992.patch, 0002-YARN-3992.patch, 
> 0003-YARN-3992.patch
>
>
> {code}
> java.lang.AssertionError: expected:<7> but was:<5>
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.failNotEquals(Assert.java:743)
>   at org.junit.Assert.assertEquals(Assert.java:118)
>   at org.junit.Assert.assertEquals(Assert.java:555)
>   at org.junit.Assert.assertEquals(Assert.java:542)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.TestApplicationPriority.testApplicationPriorityAllocation(TestApplicationPriority.java:182)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun Suresh updated YARN-3961:
--
Summary: Expose pending, running and reserved containers of a queue in REST 
api and yarn top  (was: Expose queue container information (pending, running, 
reserved) in REST api and yarn top)

> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun Suresh updated YARN-3961:
--
Affects Version/s: 2.7.2

> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Affects Versions: 2.7.2
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659551#comment-14659551
 ] 

Arun Suresh commented on YARN-3961:
---

+1, Thanks for the patch [~adhoot]
Committing this shortly..

> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Affects Versions: 2.7.2
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun Suresh updated YARN-3961:
--
Fix Version/s: 2.8.0

> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Affects Versions: 2.7.2
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Arun Suresh (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659552#comment-14659552
 ] 

Arun Suresh commented on YARN-3961:
---

Committed to trunk and branch-2

> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Affects Versions: 2.7.2
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3961) Expose pending, running and reserved containers of a queue in REST api and yarn top

2015-08-05 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14659559#comment-14659559
 ] 

Hudson commented on YARN-3961:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8271 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8271/])
YARN-3961. Expose pending, running and reserved containers of a queue in REST 
api and yarn top (adhoot via asuresh) (Arun Suresh: rev 
154c9d2e421390c808dea3b2826cdaae492ea337)
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/FairSchedulerQueueInfo.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesCapacitySched.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/records/QueueStatistics.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/AbstractCSQueue.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/api/records/impl/pb/QueueStatisticsPBImpl.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/proto/yarn_protos.proto
* hadoop-yarn-project/CHANGES.txt
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/TopCLI.java
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/CapacitySchedulerQueueInfo.java


> Expose pending, running and reserved containers of a queue in REST api and 
> yarn top
> ---
>
> Key: YARN-3961
> URL: https://issues.apache.org/jira/browse/YARN-3961
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacityscheduler, fairscheduler, webapp
>Affects Versions: 2.7.2
>Reporter: Anubhav Dhoot
>Assignee: Anubhav Dhoot
> Fix For: 2.8.0
>
> Attachments: Screen Shot 2015-07-22 at 6.17.38 PM.png, Screen Shot 
> 2015-07-22 at 6.19.31 PM.png, Screen Shot 2015-07-22 at 6.28.05 PM.png, 
> YARN-3961.001.patch, YARN-3961.001.patch, YARN-3961.002.patch
>
>
> It would be nice to expose container (allocated, pending, reserved) 
> information in the rest API and in yarn top tool



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)