[jira] [Updated] (YARN-6061) Add a customized uncaughtexceptionhandler for critical threads in RM

2017-02-09 Thread Yufei Gu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yufei Gu updated YARN-6061:
---
Attachment: YARN-6061.009.patch

> Add a customized uncaughtexceptionhandler for critical threads in RM
> 
>
> Key: YARN-6061
> URL: https://issues.apache.org/jira/browse/YARN-6061
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6061.001.patch, YARN-6061.002.patch, 
> YARN-6061.003.patch, YARN-6061.004.patch, YARN-6061.005.patch, 
> YARN-6061.006.patch, YARN-6061.007.patch, YARN-6061.008.patch, 
> YARN-6061.009.patch
>
>
> There are several threads in fair scheduler. The thread will quit when there 
> is a runtime exception inside it. We should bring down the RM when that 
> happens. Otherwise, there may be some weird behavior in RM. 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6163) FS Preemption is a trickle for severely starved applications

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860857#comment-15860857
 ] 

Hadoop QA commented on YARN-6163:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
23s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 3 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m  
4s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 15m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m  
6s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
52s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
41s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
59s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
10s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  6m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  6m 
45s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 52s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn: The patch 
generated 5 new + 104 unchanged - 1 fixed = 109 total (was 105) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
24s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
56s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  2m 
27s{color} | {color:green} hadoop-yarn-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 39m 42s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
34s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 97m 18s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.yarn.server.resourcemanager.TestRMRestart |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6163 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851992/yarn-6163-1.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 51c36a03a66d 3.13.0-107-generic #154-Ubuntu SMP Tue Dec 20 
09:57:27 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 08f9397 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-YARN-Build/14877/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14877/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
|  Test Results | 

[jira] [Commented] (YARN-6151) FS preemption does not consider child queues over fairshare if the parent is under

2017-02-09 Thread Yufei Gu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860856#comment-15860856
 ] 

Yufei Gu commented on YARN-6151:


Thanks [~kasha] for the review and commit. 

> FS preemption does not consider child queues over fairshare if the parent is 
> under
> --
>
> Key: YARN-6151
> URL: https://issues.apache.org/jira/browse/YARN-6151
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.8.0
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Fix For: 2.8.0
>
> Attachments: YARN-6151.branch-2.8.001.patch, 
> YARN-6151.branch-2.8.002.patch, YARN-6151.branch-2.8.003.patch
>
>
> This is preemption bug happens before 2.8.0, which also described in 
> YARN-3405.
> Queue hierarchy described as below:
> {noformat}
>   root
>/ \
>queue-1  queue-2   
>   /  \
> queue-1-1 queue-1-2
> {noformat}
> Assume cluster resource is 100 and all queues have same weights.
> # queue-1-1 and queue-2 has apps. Each get 50 usage and 50 fairshare. 
> # When queue-1-2 is active, supposedly it will preempt 25 from queue-1-1, but 
> this doesn't happen because preemption happens top-down, queue-2 could be the 
> preemption candidate as long as queue-2 is less needy than queue-1, and 
> queue-2 doesn't exceed the fair share which means preemption won't happen.
> We need to filter out queue-2 since it isn't a valid candidate.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6163) FS Preemption is a trickle for severely starved applications

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860854#comment-15860854
 ] 

Hadoop QA commented on YARN-6163:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
21s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 3 new or modified test 
files. {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
11s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  9m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
50s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
20s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
41s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
42s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
12s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
12s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  8m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  8m 
20s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 51s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn: The patch 
generated 5 new + 106 unchanged - 1 fixed = 111 total (was 107) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  3m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  2m 
53s{color} | {color:green} hadoop-yarn-common in the patch passed. {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 42m 25s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
31s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}103m  2s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.yarn.server.resourcemanager.TestRMRestart |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6163 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851992/yarn-6163-1.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux fd0282598150 3.13.0-105-generic #152-Ubuntu SMP Fri Dec 2 
15:37:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 08f9397 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-YARN-Build/14876/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14876/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
|  Test Results | 

[jira] [Commented] (YARN-5889) Improve and refactor user-limit calculation in capacity scheduler

2017-02-09 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860851#comment-15860851
 ] 

Sunil G commented on YARN-5889:
---

Thanks for thorough reviews and commit [~leftnoteasy] , and thanks for the 
detailed review from [~eepayne] and [~jlowe]. Really appreciate the same.
I am working on user-limit preemption poc based on this patch and will shortly 
upload in preemption jira.

> Improve and refactor user-limit calculation in capacity scheduler
> -
>
> Key: YARN-5889
> URL: https://issues.apache.org/jira/browse/YARN-5889
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacity scheduler
>Reporter: Sunil G
>Assignee: Sunil G
> Fix For: 3.0.0-alpha3
>
> Attachments: YARN-5889.0001.patch, 
> YARN-5889.0001.suggested.patchnotes, YARN-5889.0002.patch, 
> YARN-5889.0003.patch, YARN-5889.0004.patch, YARN-5889.0005.patch, 
> YARN-5889.0006.patch, YARN-5889.0007.patch, YARN-5889.0008.patch, 
> YARN-5889.0009.patch, YARN-5889.0010.patch, YARN-5889.v0.patch, 
> YARN-5889.v1.patch, YARN-5889.v2.patch
>
>
> Currently user-limit is computed during every heartbeat allocation cycle with 
> a write lock. To improve performance, this tickets is focussing on moving 
> user-limit calculation out of heartbeat allocation flow.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-6170) TimelineReaderServer should wait to join with HttpServer2

2017-02-09 Thread Sangjin Lee (JIRA)
Sangjin Lee created YARN-6170:
-

 Summary: TimelineReaderServer should wait to join with HttpServer2
 Key: YARN-6170
 URL: https://issues.apache.org/jira/browse/YARN-6170
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: timelinereader
Affects Versions: YARN-5355
Reporter: Sangjin Lee
Assignee: Sangjin Lee
Priority: Minor


While I was backporting YARN-5355-branch-2 to a 2.6.0-based code branch, I 
noticed that the timeline reader daemon would promptly shut down upon start. It 
turns out that in the 2.6.0 code line at least there are only daemon threads 
left once the main method returns. That causes the JVM to shut down.

The right pattern to start an embedded jetty web server is to call 
{{Server.start()}} followed by {{Server.join()}}. That way, the server stays up 
reliably no matter what other threads get created.

It works on YARN-5355 only because there *happens* to be one other non-daemon 
thread. We should add the {{join()}} call to be always correct.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6027) Improve /flows API for more flexible filters fromid, collapse, userid

2017-02-09 Thread Rohith Sharma K S (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860820#comment-15860820
 ] 

Rohith Sharma K S commented on YARN-6027:
-

Thanks [~sjlee0] for summarizing! Could you comment more about the reason for 
not agreeing collapse filter? (Though I was there in call, I would request 
other folks to add it).

I will create a new JIRA for discussion more on collapse filter support. 

> Improve /flows API for more flexible filters fromid, collapse, userid
> -
>
> Key: YARN-6027
> URL: https://issues.apache.org/jira/browse/YARN-6027
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Rohith Sharma K S
>Assignee: Rohith Sharma K S
>  Labels: yarn-5355-merge-blocker
> Attachments: YARN-6027-YARN-5355.0001.patch
>
>
> In YARN-5585 , fromId is supported for retrieving entities. We need similar 
> filter for flows/flowRun apps and flow run and flow as well. 
> Along with supporting fromId, this JIRA should also discuss following points
> * Should we throw an exception for entities/entity retrieval if duplicates 
> found?
> * TimelieEntity :
> ** Should equals method also check for idPrefix?
> ** Does idPrefix is part of identifiers?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5271) ATS client doesn't work with Jersey 2 on the classpath

2017-02-09 Thread Weiwei Yang (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5271?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860760#comment-15860760
 ] 

Weiwei Yang commented on YARN-5271:
---

Thanks [~gtCarrera9], hi [~jojochuang] what's your thought on this?

> ATS client doesn't work with Jersey 2 on the classpath
> --
>
> Key: YARN-5271
> URL: https://issues.apache.org/jira/browse/YARN-5271
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: client, timelineserver
>Affects Versions: 2.7.2
>Reporter: Steve Loughran
>Assignee: Weiwei Yang
>  Labels: oct16-medium
> Fix For: 2.8.0, 3.0.0-alpha2
>
> Attachments: YARN-5271.01.patch, YARN-5271.02.patch, 
> YARN-5271.branch-2.01.patch, YARN-5271-branch-2.8.01.patch
>
>
> see SPARK-15343 : once Jersey 2 is on the CP, you can't instantiate a 
> timeline client, *even if the server is an ATS1.5 server and publishing is 
> via the FS*



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6163) FS Preemption is a trickle for severely starved applications

2017-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860750#comment-15860750
 ] 

Karthik Kambatla commented on YARN-6163:


Patch that considers multiple RRs for an app. Tracking individual containers 
was too complicated for little gain. 

> FS Preemption is a trickle for severely starved applications
> 
>
> Key: YARN-6163
> URL: https://issues.apache.org/jira/browse/YARN-6163
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: fairscheduler
>Affects Versions: 2.9.0
>Reporter: Karthik Kambatla
>Assignee: Karthik Kambatla
> Attachments: yarn-6163-1.patch
>
>
> With current logic, only one RR is considered per each instance of marking an 
> application starved. This marking happens only on the update call that runs 
> every 500ms.  Due to this, an application that is severely starved takes 
> forever to reach fairshare based on preemptions.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6163) FS Preemption is a trickle for severely starved applications

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6163?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6163:
---
Attachment: yarn-6163-1.patch

> FS Preemption is a trickle for severely starved applications
> 
>
> Key: YARN-6163
> URL: https://issues.apache.org/jira/browse/YARN-6163
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: fairscheduler
>Affects Versions: 2.9.0
>Reporter: Karthik Kambatla
>Assignee: Karthik Kambatla
> Attachments: yarn-6163-1.patch
>
>
> With current logic, only one RR is considered per each instance of marking an 
> application starved. This marking happens only on the update call that runs 
> every 500ms.  Due to this, an application that is severely starved takes 
> forever to reach fairshare based on preemptions.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5703) ReservationAgents are not correctly configured

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860746#comment-15860746
 ] 

Hadoop QA commented on YARN-5703:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
19s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 2 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
56s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
36s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
16s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
5s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
33s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
33s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 21s{color} | {color:orange} 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:
 The patch generated 14 new + 94 unchanged - 0 fixed = 108 total (was 94) 
{color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 18 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
15s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
20s{color} | {color:red} 
hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager
 generated 3 new + 908 unchanged - 0 fixed = 911 total (was 908) {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 41m 52s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
20s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 65m 10s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.resourcemanager.scheduler.fair.TestFSAppStarvation |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-5703 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851427/YARN-5703.004.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux dc50c9155cd0 3.13.0-105-generic #152-Ubuntu SMP Fri Dec 2 
15:37:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 08f9397 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-YARN-Build/14875/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-YARN-Build/14875/artifact/patchprocess/whitespace-eol.txt
 |
| javadoc | 
https://builds.apache.org/job/PreCommit-YARN-Build/14875/artifact/patchprocess/diff-javadoc-javadoc-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
| unit | 

[jira] [Commented] (YARN-6141) ppc64le on Linux doesn't trigger __linux get_executable codepath

2017-02-09 Thread Sonia Garudi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860687#comment-15860687
 ] 

Sonia Garudi commented on YARN-6141:


On upgrading the version of cmake to >3.1 the build is successful. For cmake 
version < 3.1, the C standard is being set to c99 in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/CMakeLists.txt:

{code:borderStyle=solid}
if (CMAKE_VERSION VERSION_LESS "3.1")
  # subset of CMAKE__COMPILER_ID
  # 
https://cmake.org/cmake/help/v3.0/variable/CMAKE_LANG_COMPILER_ID.html
  if (CMAKE_C_COMPILER_ID STREQUAL "GNU" OR
  CMAKE_C_COMPILER_ID STREQUAL "Clang" OR
  CMAKE_C_COMPILER_ID STREQUAL "AppleClang")
set (CMAKE_C_FLAGS "-std=c99 -Wall -pedantic-errors 
${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "Intel")
set (CMAKE_C_FLAGS "-std=c99 -Wall ${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "SunPro")
set (CMAKE_C_FLAGS "-xc99 ${CMAKE_C_FLAGS}")
  endif ()
else ()
  set (CMAKE_C_STANDARD 99)
endif ()
{code}


> ppc64le on Linux doesn't trigger __linux get_executable codepath
> 
>
> Key: YARN-6141
> URL: https://issues.apache.org/jira/browse/YARN-6141
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 3.0.0-alpha3
> Environment: $ uname -a
> Linux f8eef0f055cf 3.16.0-30-generic #40~14.04.1-Ubuntu SMP Thu Jan 15 
> 17:42:36 UTC 2015 ppc64le ppc64le ppc64le GNU/Linux
>Reporter: Sonia Garudi
>  Labels: ppc64le
> Attachments: YARN-6141.diff
>
>
> On ppc64le architecture, the build fails in the 'Hadoop YARN NodeManager' 
> project with the below error:
> Cannot safely determine executable path with a relative HADOOP_CONF_DIR on 
> this operating system.
> [WARNING]  #error Cannot safely determine executable path with a relative 
> HADOOP_CONF_DIR on this operating system.
> [WARNING]   ^
> [WARNING] make[2]: *** 
> [CMakeFiles/container.dir/main/native/container-executor/impl/get_executable.c.o]
>  Error 1
> [WARNING] make[2]: *** Waiting for unfinished jobs
> [WARNING] make[1]: *** [CMakeFiles/container.dir/all] Error 2
> [WARNING] make: *** [all] Error 2
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> Cmake version used :
> $ /usr/bin/cmake --version
> cmake version 2.8.12.2



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6151) FS preemption does not consider child queues over fairshare if the parent is under

2017-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860681#comment-15860681
 ] 

Karthik Kambatla commented on YARN-6151:


+1. Just committed this to branch-2.8. 

> FS preemption does not consider child queues over fairshare if the parent is 
> under
> --
>
> Key: YARN-6151
> URL: https://issues.apache.org/jira/browse/YARN-6151
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.8.0
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6151.branch-2.8.001.patch, 
> YARN-6151.branch-2.8.002.patch, YARN-6151.branch-2.8.003.patch
>
>
> This is preemption bug happens before 2.8.0, which also described in 
> YARN-3405.
> Queue hierarchy described as below:
> {noformat}
>   root
>/ \
>queue-1  queue-2   
>   /  \
> queue-1-1 queue-1-2
> {noformat}
> Assume cluster resource is 100 and all queues have same weights.
> # queue-1-1 and queue-2 has apps. Each get 50 usage and 50 fairshare. 
> # When queue-1-2 is active, supposedly it will preempt 25 from queue-1-1, but 
> this doesn't happen because preemption happens top-down, queue-2 could be the 
> preemption candidate as long as queue-2 is less needy than queue-1, and 
> queue-2 doesn't exceed the fair share which means preemption won't happen.
> We need to filter out queue-2 since it isn't a valid candidate.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6151) FS preemption does not consider child queues over fairshare if the parent is under

2017-02-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860682#comment-15860682
 ] 

Karthik Kambatla commented on YARN-6151:


Thanks for fixing this, Yufei. 

> FS preemption does not consider child queues over fairshare if the parent is 
> under
> --
>
> Key: YARN-6151
> URL: https://issues.apache.org/jira/browse/YARN-6151
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.8.0
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6151.branch-2.8.001.patch, 
> YARN-6151.branch-2.8.002.patch, YARN-6151.branch-2.8.003.patch
>
>
> This is preemption bug happens before 2.8.0, which also described in 
> YARN-3405.
> Queue hierarchy described as below:
> {noformat}
>   root
>/ \
>queue-1  queue-2   
>   /  \
> queue-1-1 queue-1-2
> {noformat}
> Assume cluster resource is 100 and all queues have same weights.
> # queue-1-1 and queue-2 has apps. Each get 50 usage and 50 fairshare. 
> # When queue-1-2 is active, supposedly it will preempt 25 from queue-1-1, but 
> this doesn't happen because preemption happens top-down, queue-2 could be the 
> preemption candidate as long as queue-2 is less needy than queue-1, and 
> queue-2 doesn't exceed the fair share which means preemption won't happen.
> We need to filter out queue-2 since it isn't a valid candidate.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Issue Comment Deleted] (YARN-6141) ppc64le on Linux doesn't trigger __linux get_executable codepath

2017-02-09 Thread Sonia Garudi (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sonia Garudi updated YARN-6141:
---
Comment: was deleted

(was: On upgrading the version of cmake to >3.1 the build is successful. For 
cmake version < 3.1, the C standard is being set to c99 in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/CMakeLists.txt:

if (CMAKE_VERSION VERSION_LESS "3.1")
  # subset of CMAKE__COMPILER_ID
  # 
https://cmake.org/cmake/help/v3.0/variable/CMAKE_LANG_COMPILER_ID.html
  if (CMAKE_C_COMPILER_ID STREQUAL "GNU" OR
  CMAKE_C_COMPILER_ID STREQUAL "Clang" OR
  CMAKE_C_COMPILER_ID STREQUAL "AppleClang")
set (CMAKE_C_FLAGS "-std=c99 -Wall -pedantic-errors 
${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "Intel")
set (CMAKE_C_FLAGS "-std=c99 -Wall ${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "SunPro")
set (CMAKE_C_FLAGS "-xc99 ${CMAKE_C_FLAGS}")
  endif ()
else ()
  set (CMAKE_C_STANDARD 99)
endif ()
)

> ppc64le on Linux doesn't trigger __linux get_executable codepath
> 
>
> Key: YARN-6141
> URL: https://issues.apache.org/jira/browse/YARN-6141
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 3.0.0-alpha3
> Environment: $ uname -a
> Linux f8eef0f055cf 3.16.0-30-generic #40~14.04.1-Ubuntu SMP Thu Jan 15 
> 17:42:36 UTC 2015 ppc64le ppc64le ppc64le GNU/Linux
>Reporter: Sonia Garudi
>  Labels: ppc64le
> Attachments: YARN-6141.diff
>
>
> On ppc64le architecture, the build fails in the 'Hadoop YARN NodeManager' 
> project with the below error:
> Cannot safely determine executable path with a relative HADOOP_CONF_DIR on 
> this operating system.
> [WARNING]  #error Cannot safely determine executable path with a relative 
> HADOOP_CONF_DIR on this operating system.
> [WARNING]   ^
> [WARNING] make[2]: *** 
> [CMakeFiles/container.dir/main/native/container-executor/impl/get_executable.c.o]
>  Error 1
> [WARNING] make[2]: *** Waiting for unfinished jobs
> [WARNING] make[1]: *** [CMakeFiles/container.dir/all] Error 2
> [WARNING] make: *** [all] Error 2
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> Cmake version used :
> $ /usr/bin/cmake --version
> cmake version 2.8.12.2



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6151) FS preemption does not consider child queues over fairshare if the parent is under

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6151?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6151:
---
Summary: FS preemption does not consider child queues over fairshare if the 
parent is under  (was: FS Preemption doesn't filter out queues which cannot be 
preempted)

> FS preemption does not consider child queues over fairshare if the parent is 
> under
> --
>
> Key: YARN-6151
> URL: https://issues.apache.org/jira/browse/YARN-6151
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.8.0
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6151.branch-2.8.001.patch, 
> YARN-6151.branch-2.8.002.patch, YARN-6151.branch-2.8.003.patch
>
>
> This is preemption bug happens before 2.8.0, which also described in 
> YARN-3405.
> Queue hierarchy described as below:
> {noformat}
>   root
>/ \
>queue-1  queue-2   
>   /  \
> queue-1-1 queue-1-2
> {noformat}
> Assume cluster resource is 100 and all queues have same weights.
> # queue-1-1 and queue-2 has apps. Each get 50 usage and 50 fairshare. 
> # When queue-1-2 is active, supposedly it will preempt 25 from queue-1-1, but 
> this doesn't happen because preemption happens top-down, queue-2 could be the 
> preemption candidate as long as queue-2 is less needy than queue-1, and 
> queue-2 doesn't exceed the fair share which means preemption won't happen.
> We need to filter out queue-2 since it isn't a valid candidate.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6141) ppc64le on Linux doesn't trigger __linux get_executable codepath

2017-02-09 Thread Sonia Garudi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6141?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860678#comment-15860678
 ] 

Sonia Garudi commented on YARN-6141:


On upgrading the version of cmake to >3.1 the build is successful. For cmake 
version < 3.1, the C standard is being set to c99 in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/CMakeLists.txt:

if (CMAKE_VERSION VERSION_LESS "3.1")
  # subset of CMAKE__COMPILER_ID
  # 
https://cmake.org/cmake/help/v3.0/variable/CMAKE_LANG_COMPILER_ID.html
  if (CMAKE_C_COMPILER_ID STREQUAL "GNU" OR
  CMAKE_C_COMPILER_ID STREQUAL "Clang" OR
  CMAKE_C_COMPILER_ID STREQUAL "AppleClang")
set (CMAKE_C_FLAGS "-std=c99 -Wall -pedantic-errors 
${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "Intel")
set (CMAKE_C_FLAGS "-std=c99 -Wall ${CMAKE_C_FLAGS}")
  elseif (CMAKE_C_COMPILER_ID STREQUAL "SunPro")
set (CMAKE_C_FLAGS "-xc99 ${CMAKE_C_FLAGS}")
  endif ()
else ()
  set (CMAKE_C_STANDARD 99)
endif ()


> ppc64le on Linux doesn't trigger __linux get_executable codepath
> 
>
> Key: YARN-6141
> URL: https://issues.apache.org/jira/browse/YARN-6141
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 3.0.0-alpha3
> Environment: $ uname -a
> Linux f8eef0f055cf 3.16.0-30-generic #40~14.04.1-Ubuntu SMP Thu Jan 15 
> 17:42:36 UTC 2015 ppc64le ppc64le ppc64le GNU/Linux
>Reporter: Sonia Garudi
>  Labels: ppc64le
> Attachments: YARN-6141.diff
>
>
> On ppc64le architecture, the build fails in the 'Hadoop YARN NodeManager' 
> project with the below error:
> Cannot safely determine executable path with a relative HADOOP_CONF_DIR on 
> this operating system.
> [WARNING]  #error Cannot safely determine executable path with a relative 
> HADOOP_CONF_DIR on this operating system.
> [WARNING]   ^
> [WARNING] make[2]: *** 
> [CMakeFiles/container.dir/main/native/container-executor/impl/get_executable.c.o]
>  Error 1
> [WARNING] make[2]: *** Waiting for unfinished jobs
> [WARNING] make[1]: *** [CMakeFiles/container.dir/all] Error 2
> [WARNING] make: *** [all] Error 2
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> Cmake version used :
> $ /usr/bin/cmake --version
> cmake version 2.8.12.2



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6156) AM blacklistdisableThreshold consider node partition count

2017-02-09 Thread Bibin A Chundatt (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6156?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bibin A Chundatt updated YARN-6156:
---
Issue Type: Sub-task  (was: Bug)
Parent: YARN-4576

> AM blacklistdisableThreshold  consider node partition count
> ---
>
> Key: YARN-6156
> URL: https://issues.apache.org/jira/browse/YARN-6156
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Bibin A Chundatt
>Assignee: Bibin A Chundatt
>




--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6061) Add a customized uncaughtexceptionhandler for critical threads in RM

2017-02-09 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6061?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860658#comment-15860658
 ] 

ASF GitHub Bot commented on YARN-6061:
--

Github user kambatla commented on a diff in the pull request:

https://github.com/apache/hadoop/pull/182#discussion_r100468005
  
--- Diff: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMCriticalThreadUncaughtExceptionHandler.java
 ---
@@ -0,0 +1,60 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.hadoop.yarn.server.resourcemanager;
+
+import java.lang.Thread.UncaughtExceptionHandler;
+
+import org.apache.commons.logging.Log;
+import org.apache.commons.logging.LogFactory;
+import org.apache.hadoop.classification.InterfaceAudience.Public;
+import org.apache.hadoop.classification.InterfaceStability.Evolving;
+import org.apache.hadoop.yarn.conf.HAUtil;
+
+/**
+ * This class either shuts down {@link ResourceManager} or transitions the
+ * {@link ResourceManager} to standby state if a critical thread throws an
+ * uncaught exception. It is intended to be installed by calling
+ * {@code setUncaughtExceptionHandler(Thread.UncaughtExceptionHandler)}
+ * in the thread entry point or after creation of threads.
+ */
+@Public
--- End diff --

We don't need to expose this outside YARN at all. This should be @Private. 
Let us remove @Evolving altogether. 


> Add a customized uncaughtexceptionhandler for critical threads in RM
> 
>
> Key: YARN-6061
> URL: https://issues.apache.org/jira/browse/YARN-6061
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6061.001.patch, YARN-6061.002.patch, 
> YARN-6061.003.patch, YARN-6061.004.patch, YARN-6061.005.patch, 
> YARN-6061.006.patch, YARN-6061.007.patch, YARN-6061.008.patch
>
>
> There are several threads in fair scheduler. The thread will quit when there 
> is a runtime exception inside it. We should bring down the RM when that 
> happens. Otherwise, there may be some weird behavior in RM. 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-3933) Race condition when calling AbstractYarnScheduler.completedContainer.

2017-02-09 Thread Miklos Szegedi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3933?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860593#comment-15860593
 ] 

Miklos Szegedi commented on YARN-3933:
--

I could not repro the unit test issues above.

> Race condition when calling AbstractYarnScheduler.completedContainer.
> -
>
> Key: YARN-3933
> URL: https://issues.apache.org/jira/browse/YARN-3933
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.7.1
>Reporter: Lavkesh Lahngir
>Assignee: Shiwei Guo
>  Labels: oct16-medium
> Attachments: YARN-3933.001.patch, YARN-3933.002.patch, 
> YARN-3933.003.patch, YARN-3933.004.patch, YARN-3933.005.patch, 
> YARN-3933.006.patch
>
>
> In our cluster we are seeing available memory and cores being negative. 
> Initial inspection:
> Scenario no. 1: 
> In capacity scheduler the method allocateContainersToNode() checks if 
> there are excess reservation of containers for an application, and they are 
> no longer needed then it calls queue.completedContainer() which causes 
> resources being negative. And they were never assigned in the first place. 
> I am still looking through the code. Can somebody suggest how to simulate 
> excess containers assignments ?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6169) container-executor message on empty configuration file can be improved

2017-02-09 Thread Miklos Szegedi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6169?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860590#comment-15860590
 ] 

Miklos Szegedi commented on YARN-6169:
--

It would also be nice to print some additional information in the following 
case:
{code}
  if (user_info->pw_uid < min_uid && !is_whitelisted(user)) {
fprintf(LOGFILE, "Requested user %s is not whitelisted and has id %d,"
"which is below the minimum allowed %d\n", user, user_info->pw_uid, 
min_uid);
fprintf(LOGFILE, "The allowed.system.users configuration entry can be used"
" to whitelist a user");
fflush(LOGFILE);
free(user_info);
return NULL;
  }
{code}

> container-executor message on empty configuration file can be improved
> --
>
> Key: YARN-6169
> URL: https://issues.apache.org/jira/browse/YARN-6169
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Miklos Szegedi
>Priority: Trivial
>
> If the configuration file is empty, we get the following error message:
> {{Invalid configuration provided in /root/etc/hadoop/container-executor.cfg}}
> This is does not provide enough details to figure out what is the issue at 
> the first glance. We should use something like 'Empty configuration file 
> provided...'
> {code}
>   if (cfg->size == 0) {
> fprintf(ERRORFILE, "Invalid configuration provided in %s\n", file_name);
> exit(INVALID_CONFIG_FILE);
>   }
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-3933) Race condition when calling AbstractYarnScheduler.completedContainer.

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3933?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860552#comment-15860552
 ] 

Hadoop QA commented on YARN-3933:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
28s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
34s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
36s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
7s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 59m 24s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
31s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 82m 10s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.resourcemanager.TestOpportunisticContainerAllocatorAMService 
|
|   | hadoop.yarn.server.resourcemanager.TestRMRestart |
|   | hadoop.yarn.server.resourcemanager.scheduler.fair.TestFSAppStarvation |
| Timed out junit tests | 
org.apache.hadoop.yarn.server.resourcemanager.reservation.TestFairSchedulerPlanFollower
 |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-3933 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851961/YARN-3933.006.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 95cc0dc6953e 3.13.0-103-generic #150-Ubuntu SMP Thu Nov 24 
10:34:17 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 08f9397 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14874/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/14874/testReport/ |
| modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/14874/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   

[jira] [Updated] (YARN-6164) Expose maximum-am-resource-percent to Hadoop clients

2017-02-09 Thread Benson Qiu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benson Qiu updated YARN-6164:
-
Description: 
To my knowledge, there is no way to programmatically obtain 
`yarn.scheduler.capacity.maximum-am-resource-percent`.

I've tried looking at 
[YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
 [ResourceManager REST 
APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
 and 
[YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
 (via bin/yarn) 

  was:
To my knowledge, there is no way to programmatically obtain 
`yarn.scheduler.capacity.maximum-applications`.

I've tried looking at 
[YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
 [ResourceManager REST 
APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
 and 
[YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
 (via bin/yarn) 


> Expose maximum-am-resource-percent to Hadoop clients
> 
>
> Key: YARN-6164
> URL: https://issues.apache.org/jira/browse/YARN-6164
> Project: Hadoop YARN
>  Issue Type: Improvement
>Affects Versions: 2.7.2
>Reporter: Benson Qiu
> Attachments: YARN-6164.001.patch
>
>
> To my knowledge, there is no way to programmatically obtain 
> `yarn.scheduler.capacity.maximum-am-resource-percent`.
> I've tried looking at 
> [YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
>  [ResourceManager REST 
> APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
>  and 
> [YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
>  (via bin/yarn) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6164) Expose maximum-am-resource-percent to Hadoop clients

2017-02-09 Thread Benson Qiu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benson Qiu updated YARN-6164:
-
Target Version/s: 3.0.0-alpha3

> Expose maximum-am-resource-percent to Hadoop clients
> 
>
> Key: YARN-6164
> URL: https://issues.apache.org/jira/browse/YARN-6164
> Project: Hadoop YARN
>  Issue Type: Improvement
>Affects Versions: 2.7.2
>Reporter: Benson Qiu
> Attachments: YARN-6164.001.patch
>
>
> To my knowledge, there is no way to programmatically obtain 
> `yarn.scheduler.capacity.maximum-applications`.
> I've tried looking at 
> [YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
>  [ResourceManager REST 
> APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
>  and 
> [YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
>  (via bin/yarn) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6164) Expose maximum-am-resource-percent to Hadoop clients

2017-02-09 Thread Benson Qiu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benson Qiu updated YARN-6164:
-
Attachment: YARN-6164.001.patch

> Expose maximum-am-resource-percent to Hadoop clients
> 
>
> Key: YARN-6164
> URL: https://issues.apache.org/jira/browse/YARN-6164
> Project: Hadoop YARN
>  Issue Type: Improvement
>Affects Versions: 2.7.2
>Reporter: Benson Qiu
> Attachments: YARN-6164.001.patch
>
>
> To my knowledge, there is no way to programmatically obtain 
> `yarn.scheduler.capacity.maximum-applications`.
> I've tried looking at 
> [YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
>  [ResourceManager REST 
> APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
>  and 
> [YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
>  (via bin/yarn) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860481#comment-15860481
 ] 

Hadoop QA commented on YARN-6166:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
34s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
20s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
33s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
12s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 16m 
46s{color} | {color:green} hadoop-yarn-client in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
19s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 37m 35s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6166 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851955/YARN-6166.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 41e37dfb520b 3.13.0-105-generic #152-Ubuntu SMP Fri Dec 2 
15:37:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 08f9397 |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/14873/testReport/ |
| modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/14873/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> 

[jira] [Created] (YARN-6169) container-executor message on empty configuration file can be improved

2017-02-09 Thread Miklos Szegedi (JIRA)
Miklos Szegedi created YARN-6169:


 Summary: container-executor message on empty configuration file 
can be improved
 Key: YARN-6169
 URL: https://issues.apache.org/jira/browse/YARN-6169
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Miklos Szegedi
Priority: Trivial


If the configuration file is empty, we get the following error message:
{{Invalid configuration provided in /root/etc/hadoop/container-executor.cfg}}
This is does not provide enough details to figure out what is the issue at the 
first glance. We should use something like 'Empty configuration file 
provided...'
{code}
  if (cfg->size == 0) {
fprintf(ERRORFILE, "Invalid configuration provided in %s\n", file_name);
exit(INVALID_CONFIG_FILE);
  }
{code}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Miklos Szegedi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860445#comment-15860445
 ] 

Miklos Szegedi commented on YARN-6144:
--

Thank you, [~kasha].

> FairScheduler: preempted resources can become negative
> --
>
> Key: YARN-6144
> URL: https://issues.apache.org/jira/browse/YARN-6144
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: Screen Shot 2017-02-02 at 12.49.14 PM.png, 
> YARN-6144.000.patch, YARN-6144.001.patch, YARN-6144.002.patch, 
> YARN-6144.003.patch
>
>
> {{preemptContainers()}} calls {{trackContainerForPreemption()}} to collect 
> the list of containers and resources that were preempted for an application. 
> Later the list is reduced when {{containerCompleted()}} calls 
> {{untrackContainerForPreemption()}}. The bug is that the resource variable 
> {{preemptedResources}} is subtracted, not just when the container was 
> preempted but also when it has completed successfully. This causes that we 
> return an incorrect value in {{getResourceUsage()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860443#comment-15860443
 ] 

Hudson commented on YARN-6144:
--

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11229 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11229/])
YARN-6144. FairScheduler: preempted resources can become negative. (kasha: rev 
08f93978f3ec724b24a93d7ef538f158da75802f)
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSAppAttempt.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairScheduler.java


> FairScheduler: preempted resources can become negative
> --
>
> Key: YARN-6144
> URL: https://issues.apache.org/jira/browse/YARN-6144
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: Screen Shot 2017-02-02 at 12.49.14 PM.png, 
> YARN-6144.000.patch, YARN-6144.001.patch, YARN-6144.002.patch, 
> YARN-6144.003.patch
>
>
> {{preemptContainers()}} calls {{trackContainerForPreemption()}} to collect 
> the list of containers and resources that were preempted for an application. 
> Later the list is reduced when {{containerCompleted()}} calls 
> {{untrackContainerForPreemption()}}. The bug is that the resource variable 
> {{preemptedResources}} is subtracted, not just when the container was 
> preempted but also when it has completed successfully. This causes that we 
> return an incorrect value in {{getResourceUsage()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6113) re-direct NM Web Service to get container logs for finished applications

2017-02-09 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860436#comment-15860436
 ] 

Xuan Gong commented on YARN-6113:
-

[~djp] The test case failure is not related

> re-direct NM Web Service to get container logs for finished applications
> 
>
> Key: YARN-6113
> URL: https://issues.apache.org/jira/browse/YARN-6113
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Xuan Gong
>Assignee: Xuan Gong
> Attachments: YARN-6113.branch-2.v1.patch, 
> YARN-6113.branch-2.v2.patch, YARN-6113.branch-2.v3.patch, 
> YARN-6113.branch-2.v4.patch, YARN-6113.trunk.v2.patch, 
> YARN-6113.trunk.v3.patch, YARN-6113.trunk.v4.patch
>
>
> In NM web ui, when we try to get container logs for finished application, it 
> would redirect to the log server based on the configuration: 
> yarn.log.server.url. We should do the similar thing for NM WebService



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-3933) Race condition when calling AbstractYarnScheduler.completedContainer.

2017-02-09 Thread Miklos Szegedi (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Szegedi updated YARN-3933:
-
Attachment: YARN-3933.006.patch

[~guoshiwei], thank you for looking into this. I ran into the same issue in 
YARN-6158 and I had a patch there. I attach the rebased version here for your 
consideration. It is very similar to the current patch, it just addresses the 
test issue. I am also okay with 004.patch. It would be nice to get either of 
the fixes checked in. What do you think?

> Race condition when calling AbstractYarnScheduler.completedContainer.
> -
>
> Key: YARN-3933
> URL: https://issues.apache.org/jira/browse/YARN-3933
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Affects Versions: 2.7.1
>Reporter: Lavkesh Lahngir
>Assignee: Shiwei Guo
>  Labels: oct16-medium
> Attachments: YARN-3933.001.patch, YARN-3933.002.patch, 
> YARN-3933.003.patch, YARN-3933.004.patch, YARN-3933.005.patch, 
> YARN-3933.006.patch
>
>
> In our cluster we are seeing available memory and cores being negative. 
> Initial inspection:
> Scenario no. 1: 
> In capacity scheduler the method allocateContainersToNode() checks if 
> there are excess reservation of containers for an application, and they are 
> no longer needed then it calls queue.completedContainer() which causes 
> resources being negative. And they were never assigned in the first place. 
> I am still looking through the code. Can somebody suggest how to simulate 
> excess containers assignments ?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5501) Container Pooling in YARN

2017-02-09 Thread Hitesh Sharma (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860407#comment-15860407
 ] 

Hitesh Sharma commented on YARN-5501:
-

Thanks for the continued feedback, [~jlowe].

bq. Thanks for the detailed answers.  I highly recommend these get captured in 
a new revision of the design document along with diagrams to help others come 
up to speed and join the discussion.  Otherwise we have a wall of text in this 
JIRA that is essential reading for anyone understanding what is really being 
proposed.

Sure, I will do that.


bq. My thinking is the mere fact that the container finishes is indicative that 
it is ready for reuse.  Maybe there's an explicit API they call at the end of 
the task or whatever, but the same has to be done for this existing design as 
well -- we need to know when the container is ready to be reused.  The main 
difference I see in this approach is that there isn't an explicit 'pre-init' 
step where the users or admins need to premeditate what will be run.  Instead 
the first run of the app framework is the same performance it is today, but 
subsequent runs are faster since it can reused those cached containers.  Seems 
to me the most difficult part of this is coming up with an efficient container 
request protocol so YARN can know when it can reuse an old, cached container 
and when it cannot.  The existing proposal works around this by requiring the 
containers to be setup beforehand as special resource types, but that won't 
work for a general container caching approach.

Yes, once we have a way to know that the container is ready to be reused then 
we can issue a detach on it and add it to the container pool. We had a similar 
issue in our PoC where we needed to know that the container is pre-initialized 
or not (as the launched process can take some minutes to be fully ready). As 
you also said there is no protocol between YARN NM and the container to know 
what's going on, so we ended up looking for a trigger file in the pre-init 
container working directory to detect that pre-initialization is done, and the 
container can be inducted into the pool after that. We can use something 
similar to this for e.g. containers can create a trigger file and launcher 
looks for that. If found it detaches the container and inducts it into the 
pool. 

bq. They certainly are enforced, but how does the app know about the new 
constraints so they can either avoid getting shot or take advantage of the new 
space?  Simply updating the cgroup is not going to be sufficient.  Either the 
process will OOM because it slams into the new lower limit (potentially 
instantly if it is already bigger than the new limit) or it will be completely 
oblivious that it now has acres of memory that it can use.  If it tried to use 
it before it would fail, so how does it know it grew?  For example, the JVM 
can't magically do this unless the app is doing some sort of explicit off-heap 
memory management via direct buffers, etc. and is told about its memory limit.  
Simply updating the cgroup setting doesn't seem to be a sufficient 
communication channel here, so I'm curious how that's all you need to do for 
your scenario.

It is ok for our scenario because the process we pre-initialize don't do any 
work and simply initialize themselves. These processes also happen to be not 
JVM in general (and where we do use JVM I don't think we specify resource 
limits when starting the processes). When the actual AM comes around and issues 
work then the processes started a priori require resources and that time we 
adjust the cgroup or job object. Strictly speaking we aren't using resource 
resizing as it is in YARN but have our own mechanisms to update the resource 
constraints.

> Container Pooling in YARN
> -
>
> Key: YARN-5501
> URL: https://issues.apache.org/jira/browse/YARN-5501
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Arun Suresh
>Assignee: Hitesh Sharma
> Attachments: Container Pooling - one pager.pdf
>
>
> This JIRA proposes a method for reducing the container launch latency in 
> YARN. It introduces a notion of pooling *Unattached Pre-Initialized 
> Containers*.
> Proposal in brief:
> * Have a *Pre-Initialized Container Factory* service within the NM to create 
> these unattached containers.
> * The NM would then advertise these containers as special resource types 
> (this should be possible via YARN-3926).
> * When a start container request is received by the node manager for 
> launching a container requesting this specific type of resource, it will take 
> one of these unattached pre-initialized containers from the pool, and use it 
> to service the container request.
> * Once the request is complete, the pre-initialized container would be 
> released and ready to serve another 

[jira] [Updated] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6144:
---
Fix Version/s: 3.0.0-alpha2

> FairScheduler: preempted resources can become negative
> --
>
> Key: YARN-6144
> URL: https://issues.apache.org/jira/browse/YARN-6144
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: Screen Shot 2017-02-02 at 12.49.14 PM.png, 
> YARN-6144.000.patch, YARN-6144.001.patch, YARN-6144.002.patch, 
> YARN-6144.003.patch
>
>
> {{preemptContainers()}} calls {{trackContainerForPreemption()}} to collect 
> the list of containers and resources that were preempted for an application. 
> Later the list is reduced when {{containerCompleted()}} calls 
> {{untrackContainerForPreemption()}}. The bug is that the resource variable 
> {{preemptedResources}} is subtracted, not just when the container was 
> preempted but also when it has completed successfully. This causes that we 
> return an incorrect value in {{getResourceUsage()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6112) UpdateCallDuration is calculated only when debug logging is enabled

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6112:
---
Fix Version/s: 3.0.0-alpha3

> UpdateCallDuration is calculated only when debug logging is enabled
> ---
>
> Key: YARN-6112
> URL: https://issues.apache.org/jira/browse/YARN-6112
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: YARN-6112.001.patch, YARN-6112.002.patch, 
> YARN-6112.003.patch, YARN-6112.branch-2.001.patch
>
>
> In the update thread of Fair Scheduler, the 
> {{fsOpDurations.addUpdateCallDuration()}} records the duration of 
> {{update()}}, it should be independent to LOG level. YARN-4752 put the it 
> inside a {{LOG.isDebugEnabled()}} block. Not sure any particular reason to do 
> that. cc [~kasha]



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6144:
---
Fix Version/s: (was: 3.0.0-alpha2)
   3.0.0-alpha3

> FairScheduler: preempted resources can become negative
> --
>
> Key: YARN-6144
> URL: https://issues.apache.org/jira/browse/YARN-6144
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
>Priority: Blocker
> Fix For: 2.9.0, 3.0.0-alpha3
>
> Attachments: Screen Shot 2017-02-02 at 12.49.14 PM.png, 
> YARN-6144.000.patch, YARN-6144.001.patch, YARN-6144.002.patch, 
> YARN-6144.003.patch
>
>
> {{preemptContainers()}} calls {{trackContainerForPreemption()}} to collect 
> the list of containers and resources that were preempted for an application. 
> Later the list is reduced when {{containerCompleted()}} calls 
> {{untrackContainerForPreemption()}}. The bug is that the resource variable 
> {{preemptedResources}} is subtracted, not just when the container was 
> preempted but also when it has completed successfully. This causes that we 
> return an incorrect value in {{getResourceUsage()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6164) Expose maximum-am-resource-percent to Hadoop clients

2017-02-09 Thread Benson Qiu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benson Qiu updated YARN-6164:
-
Description: 
To my knowledge, there is no way to programmatically obtain 
`yarn.scheduler.capacity.maximum-applications`.

I've tried looking at 
[YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
 [ResourceManager REST 
APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
 and 
[YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
 (via bin/yarn) 

  was:
To my knowledge, there is no way to programmatically obtain 
`yarn.scheduler.capacity.maximum-applications`.

I've tried looking at 
[YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
 [ResourceManager REST 
APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
 and 
[YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
 (via bin/yarn.sh) 


> Expose maximum-am-resource-percent to Hadoop clients
> 
>
> Key: YARN-6164
> URL: https://issues.apache.org/jira/browse/YARN-6164
> Project: Hadoop YARN
>  Issue Type: Improvement
>Affects Versions: 2.7.2
>Reporter: Benson Qiu
>
> To my knowledge, there is no way to programmatically obtain 
> `yarn.scheduler.capacity.maximum-applications`.
> I've tried looking at 
> [YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
>  [ResourceManager REST 
> APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
>  and 
> [YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
>  (via bin/yarn) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-6168) Restarted RM may not inform AM about all existing containers

2017-02-09 Thread Billie Rinaldi (JIRA)
Billie Rinaldi created YARN-6168:


 Summary: Restarted RM may not inform AM about all existing 
containers
 Key: YARN-6168
 URL: https://issues.apache.org/jira/browse/YARN-6168
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Billie Rinaldi


There appears to be a race condition when an RM is restarted. I had a situation 
where the RMs and AM were down, but NMs and app containers were still running. 
When I restarted the RM, the AM restarted, registered with the RM, and received 
its list of existing containers before the NMs had reported all of their 
containers to the RM. The AM was only told about some of the app's existing 
containers.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860358#comment-15860358
 ] 

Daniel Templeton commented on YARN-6166:


LGTM.  +1

Next, time, though, it's better to keep all the patches attached to the JIRA so 
that the history is preserved.  To avoid confusion, we usually name them 
YARN-6166.001.patch, YARN-6166.002.patch, etc.

I'll commit it in a bit.

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-3666) Federation Intercepting and propagating AM-RM communications

2017-02-09 Thread Subru Krishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Subru Krishnan reassigned YARN-3666:


Assignee: Botong Huang  (was: Kishore Chaliparambil)

> Federation Intercepting and propagating AM-RM communications
> 
>
> Key: YARN-3666
> URL: https://issues.apache.org/jira/browse/YARN-3666
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, resourcemanager
>Reporter: Kishore Chaliparambil
>Assignee: Botong Huang
>
> In order, to support transparent "spanning" of jobs across sub-clusters, all 
> AM-RM communications are proxied (via YARN-2884).
> This JIRA tracks federation-specific mechanisms that decide how to 
> "split/broadcast" requests to the RMs and "merge" answers to 
> the AM.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-5531) UnmanagedAM pool manager for federating application across clusters

2017-02-09 Thread Subru Krishnan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-5531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Subru Krishnan reassigned YARN-5531:


Assignee: Botong Huang  (was: Sarvesh Sakalanaga)

> UnmanagedAM pool manager for federating application across clusters
> ---
>
> Key: YARN-5531
> URL: https://issues.apache.org/jira/browse/YARN-5531
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, resourcemanager
>Reporter: Subru Krishnan
>Assignee: Botong Huang
>
> One of the main tenets the YARN Federation is to *transparently* scale 
> applications across multiple clusters. This is achieved by running UAMs on 
> behalf of the application on other clusters. This JIRA tracks the addition of 
> a UnmanagedAM pool manager for federating application across clusters which 
> will be used the FederationInterceptor (YARN-3666) which is part of the 
> AMRMProxy pipeline introduced in YARN-2884.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6164) Expose maximum-am-resource-percent to Hadoop clients

2017-02-09 Thread Benson Qiu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6164?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860357#comment-15860357
 ] 

Benson Qiu commented on YARN-6164:
--

I built the latest version of Hadoop and it appears that 
[maxAMLimitPercentage|https://github.com/apache/hadoop/blob/f67237cbe7bc48a1b9088e990800b37529f1db2a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/QueueCapacitiesInfo.java#L52]
 has been added to the Cluster Scheduler API (/ws/v1/cluster/scheduler).

I still don't believe this config is exposed through YarnClient or YarnCLI 
though.

> Expose maximum-am-resource-percent to Hadoop clients
> 
>
> Key: YARN-6164
> URL: https://issues.apache.org/jira/browse/YARN-6164
> Project: Hadoop YARN
>  Issue Type: Improvement
>Affects Versions: 2.7.2
>Reporter: Benson Qiu
>
> To my knowledge, there is no way to programmatically obtain 
> `yarn.scheduler.capacity.maximum-applications`.
> I've tried looking at 
> [YarnClient|https://hadoop.apache.org/docs/r2.7.2/api/index.html?org/apache/hadoop/yarn/client/api/YarnClient.html],
>  [ResourceManager REST 
> APIs|https://hadoop.apache.org/docs/r2.7.0/hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html],
>  and 
> [YarnCommands|https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/YarnCommands.html]
>  (via bin/yarn.sh) 



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860349#comment-15860349
 ] 

Grant W commented on YARN-6166:
---

I've taken another crack at the patch.  Hopefully I don't continue to embarrass 
myself.

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-6167) RM option to delegate NM loss container action to AM

2017-02-09 Thread Billie Rinaldi (JIRA)
Billie Rinaldi created YARN-6167:


 Summary: RM option to delegate NM loss container action to AM
 Key: YARN-6167
 URL: https://issues.apache.org/jira/browse/YARN-6167
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: scheduler
Reporter: Billie Rinaldi
 Fix For: yarn-native-services


Currently, if the RM times out an NM, the scheduler will kill all containers 
that were running on the NM. For some applications, in the event of a temporary 
NM outage, it might be better to delegate to the AM the decision whether to 
kill the containers and request new containers from the RM.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant W updated YARN-6166:
--
Attachment: YARN-6166.patch

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant W updated YARN-6166:
--
Attachment: (was: YARN-6166.patch)

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton reassigned YARN-6166:
--

Assignee: Grant W  (was: haosdent)

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860342#comment-15860342
 ] 

Daniel Templeton commented on YARN-6166:


I assigned the JIRA to you, [~genericuser].  {{interrupt()}} isn't a static 
method, so I doubt that patch will compile.  You'll have to get the current 
thread to call {{interrupt()}} on it.  See 
https://www.ibm.com/developerworks/library/j-jtp05236/.

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: Grant W
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6056) Yarn NM using LCE shows a failure when trying to delete a non-existing dir

2017-02-09 Thread Jonathan Hung (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860337#comment-15860337
 ] 

Jonathan Hung commented on YARN-6056:
-

Thanks [~varun_saxena]! Any update on this?

> Yarn NM using LCE shows a failure when trying to delete a non-existing dir
> --
>
> Key: YARN-6056
> URL: https://issues.apache.org/jira/browse/YARN-6056
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: yarn
>Affects Versions: 2.6.4
>Reporter: Wilfred Spiegelenburg
>Assignee: Wilfred Spiegelenburg
> Attachments: YARN-6056-branch-2.6.01.patch, 
> YARN-6056-branch-2.6.1.patch
>
>
> As part of YARN-2902 the clean up of the local directories was changed to 
> ignore non existing directories and proceed with others in the list. This 
> part of the code change was not backported into branch-2.6, backporting just 
> that part now.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860332#comment-15860332
 ] 

Grant W commented on YARN-6166:
---

[~haosd...@gmail.com] got pulled in from the previous issue I cloned.  I don't 
seem to have any way to remove him from assignee.

I attached a new patch file with the interrupt call.

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant W updated YARN-6166:
--
Attachment: YARN-6166.patch

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant W updated YARN-6166:
--
Attachment: (was: YARN-6166.patch)

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (YARN-6158) FairScheduler: app usage can go to negative

2017-02-09 Thread Miklos Szegedi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860301#comment-15860301
 ] 

Miklos Szegedi edited comment on YARN-6158 at 2/9/17 10:18 PM:
---

I just noticed, I think it is the same issue as YARN-3933. This patch copies, 
what CapacityScheduler does, so it helps any refactoring in the future to merge 
some codebase.


was (Author: miklos.szeg...@cloudera.com):
I just noticed, think it is the same issue as YARN-3933. This patch copies, 
what CapacityScheduler does, so it helps any refactoring in the future to merge 
some codebase.

> FairScheduler: app usage can go to negative
> ---
>
> Key: YARN-6158
> URL: https://issues.apache.org/jira/browse/YARN-6158
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
> Attachments: YARN-6158.000.patch
>
>
> FiCaSchedulerApp.containerCompleted checks, if the container being completed 
> is in the active list:
> {code}
>   // Remove from the list of containers
>   if (null == liveContainers.remove(containerId)) {
> return false;
>   }
> {code}
> Fair scheduler should do the same, otherwise multiple different container 
> close events leave the application with negative resource usage in 
> {{queue.getMetrics().releaseResources}} and {{attemptResourceUsage.decUsed}}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860299#comment-15860299
 ] 

Hadoop QA commented on YARN-6144:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
22s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 
44s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
35s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
35s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
15s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
7s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 20s{color} | {color:orange} 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:
 The patch generated 1 new + 201 unchanged - 1 fixed = 202 total (was 202) 
{color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
11s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
6s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
17s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 39m 11s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
21s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 61m 47s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.yarn.server.resourcemanager.TestRMRestart |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6144 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851944/YARN-6144.003.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux ca6a75d720ee 3.13.0-103-generic #150-Ubuntu SMP Thu Nov 24 
10:34:17 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5fb723b |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| checkstyle | 
https://builds.apache.org/job/PreCommit-YARN-Build/14872/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14872/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/14872/testReport/ |
| modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 |
| Console output | 

[jira] [Commented] (YARN-6158) FairScheduler: app usage can go to negative

2017-02-09 Thread Miklos Szegedi (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860301#comment-15860301
 ] 

Miklos Szegedi commented on YARN-6158:
--

I just noticed, think it is the same issue as YARN-3933. This patch copies, 
what CapacityScheduler does, so it helps any refactoring in the future to merge 
some codebase.

> FairScheduler: app usage can go to negative
> ---
>
> Key: YARN-6158
> URL: https://issues.apache.org/jira/browse/YARN-6158
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
> Attachments: YARN-6158.000.patch
>
>
> FiCaSchedulerApp.containerCompleted checks, if the container being completed 
> is in the active list:
> {code}
>   // Remove from the list of containers
>   if (null == liveContainers.remove(containerId)) {
> return false;
>   }
> {code}
> Fair scheduler should do the same, otherwise multiple different container 
> close events leave the application with negative resource usage in 
> {{queue.getMetrics().releaseResources}} and {{attemptResourceUsage.decUsed}}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860277#comment-15860277
 ] 

Daniel Templeton commented on YARN-6166:


Thanks for the patch, [~haosd...@gmail.com].  Could you also add a call to 
{{Thread.interrupt()}} before the {{continue}}?

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860245#comment-15860245
 ] 

Hadoop QA commented on YARN-6166:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
23s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 16m 
 8s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
22s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
17s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
24s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
16s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
32s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
16s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
19s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
16s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
49s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
17s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 17m 25s{color} 
| {color:red} hadoop-yarn-client in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
18s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 40m 29s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.yarn.client.api.impl.TestAMRMProxy |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6166 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851942/YARN-6166.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux ede92cdad065 3.13.0-105-generic #152-Ubuntu SMP Fri Dec 2 
15:37:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5fb723b |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14871/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-client.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/14871/testReport/ |
| modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/14871/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: 

[jira] [Commented] (YARN-6112) UpdateCallDuration is calculated only when debug logging is enabled

2017-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6112?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860241#comment-15860241
 ] 

Hudson commented on YARN-6112:
--

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11228 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11228/])
YARN-6112. UpdateCallDuration is calculated only when debug logging is (kasha: 
rev 9b85053583a3560f93062b656061d11b1b9c664f)
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairScheduler.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSOpDurations.java


> UpdateCallDuration is calculated only when debug logging is enabled
> ---
>
> Key: YARN-6112
> URL: https://issues.apache.org/jira/browse/YARN-6112
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Fix For: 2.9.0
>
> Attachments: YARN-6112.001.patch, YARN-6112.002.patch, 
> YARN-6112.003.patch, YARN-6112.branch-2.001.patch
>
>
> In the update thread of Fair Scheduler, the 
> {{fsOpDurations.addUpdateCallDuration()}} records the duration of 
> {{update()}}, it should be independent to LOG level. YARN-4752 put the it 
> inside a {{LOG.isDebugEnabled()}} block. Not sure any particular reason to do 
> that. cc [~kasha]



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated YARN-6166:
---
Description: 
Logs like the following should be debug or else every legitimate stop causes 
unnecessary exception traces in the logs.

{noformat}
2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
Interrupted while waiting for queue
 java.lang.InterruptedException
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
 java:1961)
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
   at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
{noformat}

  was:
Logs like the following should be debug or else every legitimate stop causes 
unnecessary exception traces in the logs.

2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
Interrupted while waiting for queue
 java.lang.InterruptedException
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
 java:1961)
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
   at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)


> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> {noformat}
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6112) UpdateCallDuration is calculated only when debug logging is enabled

2017-02-09 Thread Yufei Gu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6112?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860212#comment-15860212
 ] 

Yufei Gu commented on YARN-6112:


Thanks [~kasha] for the review and commit.

> UpdateCallDuration is calculated only when debug logging is enabled
> ---
>
> Key: YARN-6112
> URL: https://issues.apache.org/jira/browse/YARN-6112
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Fix For: 2.9.0
>
> Attachments: YARN-6112.001.patch, YARN-6112.002.patch, 
> YARN-6112.003.patch, YARN-6112.branch-2.001.patch
>
>
> In the update thread of Fair Scheduler, the 
> {{fsOpDurations.addUpdateCallDuration()}} records the duration of 
> {{update()}}, it should be independent to LOG level. YARN-4752 put the it 
> inside a {{LOG.isDebugEnabled()}} block. Not sure any particular reason to do 
> that. cc [~kasha]



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6112) UpdateCallDuration is calculated only when debug logging is enabled

2017-02-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6112?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-6112:
---
Summary: UpdateCallDuration is calculated only when debug logging is 
enabled  (was: fsOpDurations.addUpdateCallDuration() should be independent to 
LOG level)

> UpdateCallDuration is calculated only when debug logging is enabled
> ---
>
> Key: YARN-6112
> URL: https://issues.apache.org/jira/browse/YARN-6112
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler
>Reporter: Yufei Gu
>Assignee: Yufei Gu
> Attachments: YARN-6112.001.patch, YARN-6112.002.patch, 
> YARN-6112.003.patch, YARN-6112.branch-2.001.patch
>
>
> In the update thread of Fair Scheduler, the 
> {{fsOpDurations.addUpdateCallDuration()}} records the duration of 
> {{update()}}, it should be independent to LOG level. YARN-4752 put the it 
> inside a {{LOG.isDebugEnabled()}} block. Not sure any particular reason to do 
> that. cc [~kasha]



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6144) FairScheduler: preempted resources can become negative

2017-02-09 Thread Miklos Szegedi (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Szegedi updated YARN-6144:
-
Attachment: YARN-6144.003.patch

[~kasha], that is a good point. Updating patch.

> FairScheduler: preempted resources can become negative
> --
>
> Key: YARN-6144
> URL: https://issues.apache.org/jira/browse/YARN-6144
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
>Priority: Blocker
> Attachments: Screen Shot 2017-02-02 at 12.49.14 PM.png, 
> YARN-6144.000.patch, YARN-6144.001.patch, YARN-6144.002.patch, 
> YARN-6144.003.patch
>
>
> {{preemptContainers()}} calls {{trackContainerForPreemption()}} to collect 
> the list of containers and resources that were preempted for an application. 
> Later the list is reduced when {{containerCompleted()}} calls 
> {{untrackContainerForPreemption()}}. The bug is that the resource variable 
> {{preemptedResources}} is subtracted, not just when the container was 
> preempted but also when it has completed successfully. This causes that we 
> return an incorrect value in {{getResourceUsage()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant W (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant W updated YARN-6166:
--
Attachment: YARN-6166.patch

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant W
>Assignee: haosdent
>Priority: Trivial
>  Labels: patch
> Attachments: YARN-6166.patch
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6162) Cluster Writeable APIs in RM Rest APIs page refers to APIs as alpha but they are already in use with Spark

2017-02-09 Thread Grant Sohn (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860177#comment-15860177
 ] 

Grant Sohn commented on YARN-6162:
--

I'd say the APIs are stable enough that they shouldn't be labelled alpha.  
Searching on this API reveals lots of links
from 2015 to now and some libraries/ports so it isn't really screaming bleeding 
edge anymore.


> Cluster Writeable APIs in RM Rest APIs page refers to APIs as alpha but they 
> are already in use with Spark
> --
>
> Key: YARN-6162
> URL: https://issues.apache.org/jira/browse/YARN-6162
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: documentation, site
>Reporter: Grant Sohn
>Assignee: Grant Sohn
>Priority: Trivial
>
> Excerpt with documentation that should be removed.
> {quote}
> Cluster Writeable APIs
> The setions below refer to APIs which allow to create and modify 
> applications. -These APIs are currently in alpha and may change in the 
> future.-
> Cluster New Application API
> With the New Application API, you can obtain an application-id which can then 
> be used as part of the Cluster Submit Applications API to submit 
> applications. The response also includes the maximum resource capabilities 
> available on the cluster.
> -This feature is currently in the alpha stage and may change in the future.-
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-4675) Reorganize TimelineClient and TimelineClientImpl into separate classes for ATSv1.x and ATSv2

2017-02-09 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4675?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860164#comment-15860164
 ] 

Sangjin Lee commented on YARN-4675:
---

We should be good to go once you addressed Varun's comments. Should this go to 
our feature branch (YARN-5355) or to trunk directly? I would think our feature 
branch should be the target? If so, could you kindly regenerate the patch 
against the branch?

> Reorganize TimelineClient and TimelineClientImpl into separate classes for 
> ATSv1.x and ATSv2
> 
>
> Key: YARN-4675
> URL: https://issues.apache.org/jira/browse/YARN-4675
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Naganarasimha G R
>Assignee: Naganarasimha G R
>  Labels: YARN-5355, yarn-5355-merge-blocker
> Attachments: YARN-4675.v2.002.patch, YARN-4675.v2.003.patch, 
> YARN-4675.v2.004.patch, YARN-4675.v2.005.patch, YARN-4675.v2.006.patch, 
> YARN-4675.v2.007.patch, YARN-4675-YARN-2928.v1.001.patch
>
>
> We need to reorganize TimeClientImpl into TimeClientV1Impl ,  
> TimeClientV2Impl and if required a base class, so that its clear which part 
> of the code belongs to which version and thus better maintainable.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Whiteheart updated YARN-6166:
---
Description: 
Logs like the following should be debug or else every legitimate stop causes 
unnecessary exception traces in the logs.

2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
Interrupted while waiting for queue
 java.lang.InterruptedException
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
 java:1961)
   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
   at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)

  was:
Logs like the following should be debug or else every legitimate stop causes 
unnecessary exception traces in the logs.

464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
Heartbeater interrupted
465 java.lang.InterruptedException: sleep interrupted
466   at java.lang.Thread.sleep(Native Method)
467   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
Interrupted while waiting for queue
469 java.lang.InterruptedException
470   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
 java:1961)
471   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
472   at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
473   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)


> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant Whiteheart
>Assignee: haosdent
>Priority: Trivial
>  Labels: newbie
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
>  java.lang.InterruptedException
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
>at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
>at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
>at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Whiteheart updated YARN-6166:
---
Fix Version/s: (was: 2.3.0)

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant Whiteheart
>Assignee: haosdent
>Priority: Trivial
>  Labels: newbie
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
> Heartbeater interrupted
> 465 java.lang.InterruptedException: sleep interrupted
> 466   at java.lang.Thread.sleep(Native Method)
> 467   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
> 468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
> 469 java.lang.InterruptedException
> 470   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
> 471   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
> 472   at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
> 473   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Whiteheart updated YARN-6166:
---
Affects Version/s: (was: 2.3.0)
   (was: 2.2.0)
   2.7.3

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant Whiteheart
>Assignee: haosdent
>Priority: Trivial
>  Labels: newbie
> Fix For: 2.3.0
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
> Heartbeater interrupted
> 465 java.lang.InterruptedException: sleep interrupted
> 466   at java.lang.Thread.sleep(Native Method)
> 467   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
> 468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
> 469 java.lang.InterruptedException
> 470   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
> 471   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
> 472   at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
> 473   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Grant Whiteheart updated YARN-6166:
---
Target Version/s: 2.7.4  (was: 2.3.0)

> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.7.3
>Reporter: Grant Whiteheart
>Assignee: haosdent
>Priority: Trivial
>  Labels: newbie
> Fix For: 2.3.0
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
> Heartbeater interrupted
> 465 java.lang.InterruptedException: sleep interrupted
> 466   at java.lang.Thread.sleep(Native Method)
> 467   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
> 468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
> 469 java.lang.InterruptedException
> 470   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
> 471   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
> 472   at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
> 473   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6166?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860138#comment-15860138
 ] 

Grant Whiteheart commented on YARN-6166:


YARN-1022 fixed unnecessary log messages from 
AMRMClientAsyncImpl$HeartbeatThread.run, but we still have unnecessary log 
messages from AMRMClientAsyncImpl$CallbackHandlerThread.run.

17/02/09 13:27:42 INFO impl.AMRMClientAsyncImpl: Interrupted while waiting for 
queue
java.lang.InterruptedException
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.java:2017)
at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2052)
at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442)
at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:274)


> Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run
> --
>
> Key: YARN-6166
> URL: https://issues.apache.org/jira/browse/YARN-6166
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 2.2.0, 2.3.0
>Reporter: Grant Whiteheart
>Assignee: haosdent
>Priority: Trivial
>  Labels: newbie
> Fix For: 2.3.0
>
>
> Logs like the following should be debug or else every legitimate stop causes 
> unnecessary exception traces in the logs.
> 464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
> Heartbeater interrupted
> 465 java.lang.InterruptedException: sleep interrupted
> 466   at java.lang.Thread.sleep(Native Method)
> 467   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
> 468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
> Interrupted while waiting for queue
> 469 java.lang.InterruptedException
> 470   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
>  java:1961)
> 471   at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
> 472   at 
> java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
> 473   at 
> org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-4985) Refactor the coprocessor code & other definition classes into independent packages

2017-02-09 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860136#comment-15860136
 ] 

Sangjin Lee commented on YARN-4985:
---

Thanks for refreshing my memory on this Haibo. Yes, I recall that was the main 
issue. To tease out client and server correctly, we'd need a separate (common) 
module both client and server depends on. The only way we could avoid it is if 
all the fine-grained downstream dependencies can be isolated and moved into the 
hbase-server module. Do you know if it is feasible?

If that is not feasible, maybe this refactoring would be more problematic than 
beneficial. What would we lose if we didn't do this refactoring? Also, if we 
didn't do this refactoring and kept thins as is (just timelineservice-hbase), 
do we still have the dependency issues in terms of installing the coprocessor 
(we need to follow all transitive dependencies between classes)? If so, that 
needs to be addressed no matter what.

> Refactor the coprocessor code & other definition classes into independent 
> packages
> --
>
> Key: YARN-4985
> URL: https://issues.apache.org/jira/browse/YARN-4985
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Vrushali C
>Assignee: Haibo Chen
>  Labels: YARN-5355
> Attachments: YARN-4985-YARN-5355.prelim.patch
>
>
> As part of the coprocessor deployment, we have realized that it will be much 
> cleaner to have the coprocessor code sit in a package which does not depend 
> on hadoop-yarn-server classes. It only needs hbase and other util classes.
> These util classes and tag definition related classes can be refactored into 
> their own independent "definition" class package so that making changes to 
> coprocessor code, upgrading hbase, deploying hbase on a different hadoop 
> version cluster etc all becomes operationally much easier and less error 
> prone to having different library jars etc.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-6166) Unnecessary INFO logs in AMRMClientAsyncImpl$CallbackHandlerThread.run

2017-02-09 Thread Grant Whiteheart (JIRA)
Grant Whiteheart created YARN-6166:
--

 Summary: Unnecessary INFO logs in 
AMRMClientAsyncImpl$CallbackHandlerThread.run
 Key: YARN-6166
 URL: https://issues.apache.org/jira/browse/YARN-6166
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0, 2.3.0
Reporter: Grant Whiteheart
Assignee: haosdent
Priority: Trivial
 Fix For: 2.3.0


Logs like the following should be debug or else every legitimate stop causes 
unnecessary exception traces in the logs.

464 2013-08-03 20:01:34,459 INFO [AMRM Heartbeater thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:
Heartbeater interrupted
465 java.lang.InterruptedException: sleep interrupted
466   at java.lang.Thread.sleep(Native Method)
467   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$HeartbeatThread.run(AMRMClientAsyncImpl.java:249)
468 2013-08-03 20:01:34,460 INFO [AMRM Callback Handler Thread] 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl:   
Interrupted while waiting for queue
469 java.lang.InterruptedException
470   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.reportInterruptAfterWait(AbstractQueuedSynchronizer.
 java:1961)
471   at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1996)
472   at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
473   at 
org.apache.hadoop.yarn.client.api.async.impl.AMRMClientAsyncImpl$CallbackHandlerThread.run(AMRMClientAsyncImpl.java:275)



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6027) Improve /flows API for more flexible filters fromid, collapse, userid

2017-02-09 Thread Sangjin Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860111#comment-15860111
 ] 

Sangjin Lee commented on YARN-6027:
---

We discussed this offline during our status call. Just to record what we 
agreed, can we separate the aspect of the collapse filter from this JIRA and 
focus on fromId and userId? I don't think there is an agreement on the collapse 
part, and we can still make progress with the other two filters separately. 
Thanks!

> Improve /flows API for more flexible filters fromid, collapse, userid
> -
>
> Key: YARN-6027
> URL: https://issues.apache.org/jira/browse/YARN-6027
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Rohith Sharma K S
>Assignee: Rohith Sharma K S
>  Labels: yarn-5355-merge-blocker
> Attachments: YARN-6027-YARN-5355.0001.patch
>
>
> In YARN-5585 , fromId is supported for retrieving entities. We need similar 
> filter for flows/flowRun apps and flow run and flow as well. 
> Along with supporting fromId, this JIRA should also discuss following points
> * Should we throw an exception for entities/entity retrieval if duplicates 
> found?
> * TimelieEntity :
> ** Should equals method also check for idPrefix?
> ** Does idPrefix is part of identifiers?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6162) Cluster Writeable APIs in RM Rest APIs page refers to APIs as alpha but they are already in use with Spark

2017-02-09 Thread Daniel Templeton (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6162?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860079#comment-15860079
 ] 

Daniel Templeton commented on YARN-6162:


Is the issue that the APIs are mislabeled or that Spark is using an alpha API?

> Cluster Writeable APIs in RM Rest APIs page refers to APIs as alpha but they 
> are already in use with Spark
> --
>
> Key: YARN-6162
> URL: https://issues.apache.org/jira/browse/YARN-6162
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: documentation, site
>Reporter: Grant Sohn
>Assignee: Grant Sohn
>Priority: Trivial
>
> Excerpt with documentation that should be removed.
> {quote}
> Cluster Writeable APIs
> The setions below refer to APIs which allow to create and modify 
> applications. -These APIs are currently in alpha and may change in the 
> future.-
> Cluster New Application API
> With the New Application API, you can obtain an application-id which can then 
> be used as part of the Cluster Submit Applications API to submit 
> applications. The response also includes the maximum resource capabilities 
> available on the cluster.
> -This feature is currently in the alpha stage and may change in the future.-
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6158) FairScheduler: app usage can go to negative

2017-02-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6158?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860057#comment-15860057
 ] 

Hadoop QA commented on YARN-6158:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
19s{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 14m 
12s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
26s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
40s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
16s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
12s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
37s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green}  0m 
13s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
22s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 43m 43s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
23s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 67m 37s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler |
|   | hadoop.yarn.server.resourcemanager.scheduler.fair.TestFSAppStarvation |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:a9ad5d6 |
| JIRA Issue | YARN-6158 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12851915/YARN-6158.000.patch |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux a2c1e8fd98ff 3.13.0-105-generic #152-Ubuntu SMP Fri Dec 2 
15:37:11 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / 5fb723b |
| Default Java | 1.8.0_121 |
| findbugs | v3.0.0 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/14870/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/14870/testReport/ |
| modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/14870/console |
| Powered by | Apache Yetus 0.5.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> FairScheduler: app usage can go to negative
> ---
>
> Key: YARN-6158
> 

[jira] [Commented] (YARN-4090) Make Collections.sort() more efficient in FSParentQueue.java

2017-02-09 Thread Yufei Gu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4090?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15860008#comment-15860008
 ] 

Yufei Gu commented on YARN-4090:


Thanks [~zsl2007].
I think you get it covered. Things might affect resource usage would be:
# Submit an application
# Allocate tasks
# Move app from one queue to another
# Complete containers (both apps and tasks finish)
# Preempt containers 
# Nodes are down, this might not affect it directly, need to double check. 
Thanks Ray for pointing out. 

I could miss something. Correct me if you have anything else. I guess the test 
cases need to cover these. We might not need to worry about queue deletion 
since it seems not supported. 

YARN-4691 is about same thing for FSLeafQueue. Make sense to me to combine 
YARN-4691 in this JIRA. 

> Make Collections.sort() more efficient in FSParentQueue.java
> 
>
> Key: YARN-4090
> URL: https://issues.apache.org/jira/browse/YARN-4090
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: fairscheduler
>Reporter: Xianyin Xin
>Assignee: zhangshilong
> Attachments: sampling1.jpg, sampling2.jpg, YARN-4090.001.patch, 
> YARN-4090.002.patch, YARN-4090.003.patch, YARN-4090.004.patch, 
> YARN-4090.005.patch, YARN-4090.006.patch, YARN-4090-preview.patch, 
> YARN-4090-TestResult.pdf
>
>
> Collections.sort() consumes too much time in a scheduling round.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5889) Improve and refactor user-limit calculation in capacity scheduler

2017-02-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5889?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859996#comment-15859996
 ] 

Hudson commented on YARN-5889:
--

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #11227 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/11227/])
YARN-5889. Improve and refactor user-limit calculation in Capacity (wangda: rev 
5fb723bb77722d41df6959eee23e1b0cfeb5584e)
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/dao/CapacitySchedulerLeafQueueInfo.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSParentQueue.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/monitor/capacity/FifoIntraQueuePreemptionPlugin.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/ActiveUsersManager.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/ParentQueue.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestCapacitySchedulerNodeLabelUpdate.java
* (add) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AbstractUsersManager.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestNodeLabelContainerAllocation.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityScheduler.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CapacityHeadroomProvider.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/common/fica/FiCaSchedulerApp.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/monitor/capacity/ProportionalCapacityPreemptionPolicyMockFramework.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FSLeafQueue.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestApplicationLimitsByPartition.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/SchedulerApplicationAttempt.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/CSQueue.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fifo/FifoScheduler.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/Queue.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestApplicationLimits.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/AppSchedulingInfo.java
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerApplicationAttempt.java
* (add) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/UsersManager.java
* (edit) 

[jira] [Assigned] (YARN-6165) Intra-queue preemption occurs even when preemption is turned off for a specific queue.

2017-02-09 Thread Eric Payne (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6165?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Payne reassigned YARN-6165:


Assignee: Eric Payne

> Intra-queue preemption occurs even when preemption is turned off for a 
> specific queue.
> --
>
> Key: YARN-6165
> URL: https://issues.apache.org/jira/browse/YARN-6165
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacity scheduler, scheduler preemption
>Affects Versions: 3.0.0-alpha2
>Reporter: Eric Payne
>Assignee: Eric Payne
>
> Intra-queue preemption occurs even when preemption is turned on for the whole 
> cluster ({{yarn.resourcemanager.scheduler.monitor.enable == true}}) but 
> turned off for a specific queue 
> ({{yarn.scheduler.capacity.root.queue1.disable_preemption == true}}).



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6165) Intra-queue preemption occurs even when preemption is turned off for a specific queue.

2017-02-09 Thread Eric Payne (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6165?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859987#comment-15859987
 ] 

Eric Payne commented on YARN-6165:
--

Use Case:
- Configure queues with cluster-wide preemption on, but a specific queue's 
preemption off (see above).
- Submit a job at priority 1 that fills the entire queue
- Submit a job as the same user at priority 2

Containers from the first job will be preempted when they shouldn't be.

> Intra-queue preemption occurs even when preemption is turned off for a 
> specific queue.
> --
>
> Key: YARN-6165
> URL: https://issues.apache.org/jira/browse/YARN-6165
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacity scheduler, scheduler preemption
>Affects Versions: 3.0.0-alpha2
>Reporter: Eric Payne
>
> Intra-queue preemption occurs even when preemption is turned on for the whole 
> cluster ({{yarn.resourcemanager.scheduler.monitor.enable == true}}) but 
> turned off for a specific queue 
> ({{yarn.scheduler.capacity.root.queue1.disable_preemption == true}}).



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-5889) Improve and refactor user-limit calculation in capacity scheduler

2017-02-09 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-5889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-5889:
-
Summary: Improve and refactor user-limit calculation in capacity scheduler  
(was: Improve user-limit calculation in capacity scheduler)

> Improve and refactor user-limit calculation in capacity scheduler
> -
>
> Key: YARN-5889
> URL: https://issues.apache.org/jira/browse/YARN-5889
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: capacity scheduler
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: YARN-5889.0001.patch, 
> YARN-5889.0001.suggested.patchnotes, YARN-5889.0002.patch, 
> YARN-5889.0003.patch, YARN-5889.0004.patch, YARN-5889.0005.patch, 
> YARN-5889.0006.patch, YARN-5889.0007.patch, YARN-5889.0008.patch, 
> YARN-5889.0009.patch, YARN-5889.0010.patch, YARN-5889.v0.patch, 
> YARN-5889.v1.patch, YARN-5889.v2.patch
>
>
> Currently user-limit is computed during every heartbeat allocation cycle with 
> a write lock. To improve performance, this tickets is focussing on moving 
> user-limit calculation out of heartbeat allocation flow.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-6158) FairScheduler: app usage can go to negative

2017-02-09 Thread Miklos Szegedi (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-6158?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Miklos Szegedi updated YARN-6158:
-
Attachment: YARN-6158.000.patch

Attaching patch.

> FairScheduler: app usage can go to negative
> ---
>
> Key: YARN-6158
> URL: https://issues.apache.org/jira/browse/YARN-6158
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: fairscheduler, resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Miklos Szegedi
>Assignee: Miklos Szegedi
> Attachments: YARN-6158.000.patch
>
>
> FiCaSchedulerApp.containerCompleted checks, if the container being completed 
> is in the active list:
> {code}
>   // Remove from the list of containers
>   if (null == liveContainers.remove(containerId)) {
> return false;
>   }
> {code}
> Fair scheduler should do the same, otherwise multiple different container 
> close events leave the application with negative resource usage in 
> {{queue.getMetrics().releaseResources}} and {{attemptResourceUsage.decUsed}}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-6165) Intra-queue preemption occurs even when preemption is turned off for a specific queue.

2017-02-09 Thread Eric Payne (JIRA)
Eric Payne created YARN-6165:


 Summary: Intra-queue preemption occurs even when preemption is 
turned off for a specific queue.
 Key: YARN-6165
 URL: https://issues.apache.org/jira/browse/YARN-6165
 Project: Hadoop YARN
  Issue Type: Bug
  Components: capacity scheduler, scheduler preemption
Affects Versions: 3.0.0-alpha2
Reporter: Eric Payne


Intra-queue preemption occurs even when preemption is turned on for the whole 
cluster ({{yarn.resourcemanager.scheduler.monitor.enable == true}}) but turned 
off for a specific queue 
({{yarn.scheduler.capacity.root.queue1.disable_preemption == true}}).



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-4985) Refactor the coprocessor code & other definition classes into independent packages

2017-02-09 Thread Haibo Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-4985?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859912#comment-15859912
 ] 

Haibo Chen commented on YARN-4985:
--

The dependency is not mainly because all tests are in the hbase-server module 
as they are now, but because FlowScanner needs to reference FlowRun table 
schema classes. If we were to extract another module that only includes 
table/column definitions, we can avoid the dependency from hbase-server on 
hbase-client. But hbase-client code seems quite coupled to the table schema 
code due to current code implementation (read & write methods are defined 
within the Table/Column classes).  I will try to break the current hbase-client 
module in my patch into hbase-schema and hbase-client. This way, we should not 
have the undesirable dependency any more. 

That said, I think my previous concern about loading coprocessor from HDFS is 
still valid, since we will still have hbase-server depend on hbase-schema.

> Refactor the coprocessor code & other definition classes into independent 
> packages
> --
>
> Key: YARN-4985
> URL: https://issues.apache.org/jira/browse/YARN-4985
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Vrushali C
>Assignee: Haibo Chen
>  Labels: YARN-5355
> Attachments: YARN-4985-YARN-5355.prelim.patch
>
>
> As part of the coprocessor deployment, we have realized that it will be much 
> cleaner to have the coprocessor code sit in a package which does not depend 
> on hadoop-yarn-server classes. It only needs hbase and other util classes.
> These util classes and tag definition related classes can be refactored into 
> their own independent "definition" class package so that making changes to 
> coprocessor code, upgrading hbase, deploying hbase on a different hadoop 
> version cluster etc all becomes operationally much easier and less error 
> prone to having different library jars etc.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Resolved] (YARN-5700) testAMRestartNotLostContainerCompleteMsg times out intermittently in 2.8

2017-02-09 Thread Eric Badger (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-5700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Badger resolved YARN-5700.
---
Resolution: Cannot Reproduce

Revisited this and I cannot reproduce the failure. Closing as cannot reproduce

> testAMRestartNotLostContainerCompleteMsg times out intermittently in 2.8
> 
>
> Key: YARN-5700
> URL: https://issues.apache.org/jira/browse/YARN-5700
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Eric Badger
>Assignee: Eric Badger
>
> {noformat}
> java.lang.Exception: test timed out after 3 milliseconds
>   at java.lang.Thread.sleep(Native Method)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.MockRM.waitForState(MockRM.java:301)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.MockRM.waitForState(MockRM.java:286)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.MockRM.waitForState(MockRM.java:281)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.applicationsmanager.TestAMRestart.testAMRestartNotLostContainerCompleteMsg(TestAMRestart.java:774)
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5956) Refactor ClientRMService

2017-02-09 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859779#comment-15859779
 ] 

Sunil G commented on YARN-5956:
---

Thanks [~lewuathe].

Were u mentioning about {{YARN-5956.08.patch}}. IN that patch also, 
{{verifyUserAccessForRMApp}} internally calling {{checkAccess}}, correct? Am I 
missing something.

> Refactor ClientRMService
> 
>
> Key: YARN-5956
> URL: https://issues.apache.org/jira/browse/YARN-5956
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Kai Sasaki
>Assignee: Kai Sasaki
>Priority: Minor
> Attachments: YARN-5956.01.patch, YARN-5956.02.patch, 
> YARN-5956.03.patch, YARN-5956.04.patch, YARN-5956.05.patch, 
> YARN-5956.06.patch, YARN-5956.07.patch, YARN-5956.08.patch
>
>
> Some refactoring can be done in {{ClientRMService}}.
> - Remove redundant variable declaration
> - Fill in missing javadocs
> - Proper variable access modifier
> - Fix some typos in method name and exception messages



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-5501) Container Pooling in YARN

2017-02-09 Thread Jason Lowe (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859587#comment-15859587
 ] 

Jason Lowe commented on YARN-5501:
--

Thanks for the detailed answers.  I highly recommend these get captured in a 
new revision of the design document along with diagrams to help others come up 
to speed and join the discussion.  Otherwise we have a wall of text in this 
JIRA that is essential reading for anyone understanding what is really being 
proposed.

bq. If the app can provide some hints as to when it is good to consider a 
container pre-initialized then when the container finishes we can carry out the 
required operations to go back to the pre-init state.

My thinking is the mere fact that the container finishes is indicative that it 
is ready for reuse.  Maybe there's an explicit API they call at the end of the 
task or whatever, but the same has to be done for this existing design as well 
-- we need to know when the container is ready to be reused.  The main 
difference I see in this approach is that there isn't an explicit 'pre-init' 
step where the users or admins need to premeditate what will be run.  Instead 
the first run of the app framework is the same performance it is today, but 
subsequent runs are faster since it can reused those cached containers.  Seems 
to me the most difficult part of this is coming up with an efficient container 
request protocol so YARN can know when it can reuse an old, cached container 
and when it cannot.  The existing proposal works around this by requiring the 
containers to be setup beforehand as special resource types, but that won't 
work for a general container caching approach.

bq. What you are saying makes sense, but in that case container resizing won't 
work as well.

I don't believe it does for cases where we're trying to resize something like a 
JVM.  I honestly don't know of any real-world use cases for it today, but my 
guess is that they either have a small front-end launcher that can re-launch 
something when they are told (via a completely separate, app-specific channel) 
that the resize is about to occur/has occurred, or they have explicit memory 
caching that they can trivially grow or shrink (again, completely outside of 
any help from YARN).

bq. For our scenarios resource constraints are enforced via job objects or 
cgroups so things are ok.

They certainly are enforced, but how does the app know about the new 
constraints so they can either avoid getting shot or take advantage of the new 
space?  Simply updating the cgroup is not going to be sufficient.  Either the 
process will OOM because it slams into the new lower limit (potentially 
instantly if it is already bigger than the new limit) or it will be completely 
oblivious that it now has acres of memory that it can use.  If it tried to use 
it before it would fail, so how does it know it grew?  For example, the JVM 
can't magically do this unless the app is doing some sort of explicit off-heap 
memory management via direct buffers, etc. and is told about its memory limit.  
Simply updating the cgroup setting doesn't seem to be a sufficient 
communication channel here, so I'm curious how that's all you need to do for 
your scenario.

It seems to me there needs to be a predictable sequence here, and it's not the 
same for growing and shrinking.  When we are growing the memory of a container 
it's a bit simpler.  The container needs to be signaled about the resizing 
event _after_ the resizing has occurred at the enforcement level (i.e.: 
cgroup).  This could simply be part of the code that the application runs as 
part of reusing the container.  However if we are shrinking the memory of the 
container then we need to signal the container _before_ the resizing has 
occurred _and_ we need to know when the container process(es) have had a chance 
to react to that signal and get under the new, lower limit before actually 
enforcing it (i..e: at the cgroup level).  Otherwise the only way I can see 
this working is if the container resizing isn't generalized but requires the 
container code to always shrink itself to a minimal acceptable level (i.e.: the 
smallest that will ever be requested) after running each app's task so every 
case essentially becomes the container growth scenario.


> Container Pooling in YARN
> -
>
> Key: YARN-5501
> URL: https://issues.apache.org/jira/browse/YARN-5501
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Arun Suresh
>Assignee: Hitesh Sharma
> Attachments: Container Pooling - one pager.pdf
>
>
> This JIRA proposes a method for reducing the container launch latency in 
> YARN. It introduces a notion of pooling *Unattached Pre-Initialized 
> Containers*.
> Proposal in brief:
> * Have a *Pre-Initialized Container Factory* service within the NM to 

[jira] [Commented] (YARN-5956) Refactor ClientRMService

2017-02-09 Thread Kai Sasaki (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-5956?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859464#comment-15859464
 ] 

Kai Sasaki commented on YARN-5956:
--

[~sunilg] Sorry for late. I checked redundant {{checkAccess}} call and updated. 
Failed tests seems not related to the patch. I confirmed it passed at local.
Could you check again please?

> Refactor ClientRMService
> 
>
> Key: YARN-5956
> URL: https://issues.apache.org/jira/browse/YARN-5956
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: resourcemanager
>Affects Versions: 3.0.0-alpha2
>Reporter: Kai Sasaki
>Assignee: Kai Sasaki
>Priority: Minor
> Attachments: YARN-5956.01.patch, YARN-5956.02.patch, 
> YARN-5956.03.patch, YARN-5956.04.patch, YARN-5956.05.patch, 
> YARN-5956.06.patch, YARN-5956.07.patch, YARN-5956.08.patch
>
>
> Some refactoring can be done in {{ClientRMService}}.
> - Remove redundant variable declaration
> - Fill in missing javadocs
> - Proper variable access modifier
> - Fix some typos in method name and exception messages



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6113) re-direct NM Web Service to get container logs for finished applications

2017-02-09 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6113?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859421#comment-15859421
 ] 

Junping Du commented on YARN-6113:
--

Thanks Xuan for updating the patch. Latest patch LGTM. However, does UT 
failures related to patch here? I cannot reproduce the failure w/o patch 
locally.

> re-direct NM Web Service to get container logs for finished applications
> 
>
> Key: YARN-6113
> URL: https://issues.apache.org/jira/browse/YARN-6113
> Project: Hadoop YARN
>  Issue Type: Sub-task
>Reporter: Xuan Gong
>Assignee: Xuan Gong
> Attachments: YARN-6113.branch-2.v1.patch, 
> YARN-6113.branch-2.v2.patch, YARN-6113.branch-2.v3.patch, 
> YARN-6113.branch-2.v4.patch, YARN-6113.trunk.v2.patch, 
> YARN-6113.trunk.v3.patch, YARN-6113.trunk.v4.patch
>
>
> In NM web ui, when we try to get container logs for finished application, it 
> would redirect to the log server based on the configuration: 
> yarn.log.server.url. We should do the similar thing for NM WebService



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6027) Improve /flows API for more flexible filters fromid, collapse, userid

2017-02-09 Thread Varun Saxena (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859344#comment-15859344
 ] 

Varun Saxena commented on YARN-6027:


Ok...I had forgotton that we currently do not limit flow runs which means that 
in collapse mode, if daterange is not specified, we will have to do a full 
table scan anyways. I have a JIRA for limiting flowruns. I will discuss with 
some HBase guy and check the scenario with him and handle it in that JIRA, if 
required. Basically its a choice between a full table scan of say 10 rows 
vs multiple scans with PageFilter equal to limit in each scan(or 2-3 times the 
limit value) and breaking out from the loop as soon as flow run limit is also 
reached. By default we can return 0 flow runs for flows endpoint as we already 
have a flowruns endpoint as well.
Anyways I am not a 100% sure on this and need to check with a HBase guy. We can 
get back to this on the JIRA which is with me.

> Improve /flows API for more flexible filters fromid, collapse, userid
> -
>
> Key: YARN-6027
> URL: https://issues.apache.org/jira/browse/YARN-6027
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: timelineserver
>Reporter: Rohith Sharma K S
>Assignee: Rohith Sharma K S
>  Labels: yarn-5355-merge-blocker
> Attachments: YARN-6027-YARN-5355.0001.patch
>
>
> In YARN-5585 , fromId is supported for retrieving entities. We need similar 
> filter for flows/flowRun apps and flow run and flow as well. 
> Along with supporting fromId, this JIRA should also discuss following points
> * Should we throw an exception for entities/entity retrieval if duplicates 
> found?
> * TimelieEntity :
> ** Should equals method also check for idPrefix?
> ** Does idPrefix is part of identifiers?



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-6157) Inconsistencies in verifying Max Applications

2017-02-09 Thread Varun Saxena (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-6157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15859205#comment-15859205
 ] 

Varun Saxena commented on YARN-6157:


[~Naganarasimha], the check for system max apps is made in 
CapacityScheduler#addApplication and flow for recovered apps goes via 
CapacityScheduler#addApplicationOnRecovery. So recovered apps I think will be 
excluded.
Other point is right. We probably need a counter in LeafQueue which will be 
increment when we call LeafQueue#submitApplication. We have a similar counter 
in ParentQueue too.

> Inconsistencies in verifying Max Applications
> -
>
> Key: YARN-6157
> URL: https://issues.apache.org/jira/browse/YARN-6157
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Naganarasimha G R
>Assignee: Naganarasimha G R
>
> Inconsistencies in verifying Max Applications when the max apps is reduced 
> and either HA is done ow work preserving restart is done.
> # currently Max applications across cluster should not be done for the 
> recovered apps. Seems like currently we are doing it
> #  Max applications for a queue is done @ CapacityScheduler.addApplication 
> which considers sum of Pending and running applications but we add to pending 
> applications in {{CapacityScheduler.addApplicationAttempt -> 
> LeafQueue.addApplicationAttempt}} so between these 2 checks we can activate 
> more apps than what can queue restrict.
> # During recovery of a RMApp, if applicationAttempts are not found then we 
> recover it without recovery false @ {{RMAppImpl.RMAppRecoveredTransition}}, 
> this can lead to failure of apps which were accepted earlier but attempt was 
> not yet created and HA happens when MAX app configuration (for cluster/queue) 
> is modified.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org