[jira] [Commented] (YARN-10476) Queue metrics of Unmanaged applications
[ https://issues.apache.org/jira/browse/YARN-10476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222721#comment-17222721 ] Cyrus Jackson commented on YARN-10476: -- Submitting initial patch for unit tests. > Queue metrics of Unmanaged applications > > > Key: YARN-10476 > URL: https://issues.apache.org/jira/browse/YARN-10476 > Project: Hadoop YARN > Issue Type: Improvement > Components: resourcemanager >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Minor > Attachments: YARN-10476.001.patch > > > Right now we do not have separate metrics unmanaged applications. All > application metrics come as part of Queue (Managed and UnManaged), This Jira > aims to show them separately. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10476) Queue metrics of Unmanaged applications
[ https://issues.apache.org/jira/browse/YARN-10476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10476: - Attachment: YARN-10476.001.patch > Queue metrics of Unmanaged applications > > > Key: YARN-10476 > URL: https://issues.apache.org/jira/browse/YARN-10476 > Project: Hadoop YARN > Issue Type: Improvement > Components: resourcemanager >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Minor > Attachments: YARN-10476.001.patch > > > Right now we do not have separate metrics unmanaged applications. All > application metrics come as part of Queue (Managed and UnManaged), This Jira > aims to show them separately. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10476) Queue metrics of Unmanaged applications
[ https://issues.apache.org/jira/browse/YARN-10476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10476: - Issue Type: Improvement (was: Bug) > Queue metrics of Unmanaged applications > > > Key: YARN-10476 > URL: https://issues.apache.org/jira/browse/YARN-10476 > Project: Hadoop YARN > Issue Type: Improvement > Components: resourcemanager >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Minor > > Right now we do not have separate metrics unmanaged applications. All > application metrics come as part of Queue (Managed and UnManaged), This Jira > aims to show them separately. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222615#comment-17222615 ] Hadoop QA commented on YARN-10471: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 14s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 1s{color} | | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} markdownlint {color} | {color:blue} 0m 0s{color} | | {color:blue} markdownlint was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 25s{color} | | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 23m 53s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 42s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 41s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 22s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 56s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 21m 1s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 35s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 47s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 25s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 25s{color} | | {color:blue} branch/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site no findbugs output file (findbugsXml.xml) {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 24s{color} | | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 6s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 1s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 9m 1s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 7s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 7s{color} | | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} blanks {color} | {color:red} 0m 0s{color} | [/blanks-tabs.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/268/artifact/out/blanks-tabs.txt] | {color:red} The patch 1 line(s) with tabs. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 20s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 51s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 1s{color} | | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 12s{color} | | {color:green} patch has no errors when building and testing our client artifacts. {color
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222565#comment-17222565 ] Hadoop QA commented on YARN-10471: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 16s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} markdownlint {color} | {color:blue} 0m 0s{color} | | {color:blue} markdownlint was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 37s{color} | | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 21m 58s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 35s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 1s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 22s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 57s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 21m 2s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 36s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 2m 43s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 25s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 25s{color} | | {color:blue} branch/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site no findbugs output file (findbugsXml.xml) {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 23s{color} | | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 10s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 24s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 9m 24s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 15s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 9m 15s{color} | | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} blanks {color} | {color:red} 0m 0s{color} | [/blanks-eol.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/267/artifact/out/blanks-eol.txt] | {color:red} The patch has 1 line(s) that end in blanks. Use git apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply {color} | | {color:red}-1{color} | {color:red} blanks {color} | {color:red} 0m 0s{color} | [/blanks-tabs.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/267/artifact/out/blanks-tabs.txt] | {color:red} The patch 1 line(s) with tabs. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 24s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 56s{color} | | {color:green} the patch passed {color} | | {c
[jira] [Commented] (YARN-10475) Scale RM-NM heartbeat interval based on node utilization
[ https://issues.apache.org/jira/browse/YARN-10475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222517#comment-17222517 ] Hadoop QA commented on YARN-10475: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 54s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 1s{color} | | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 11m 41s{color} | | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 57s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 20m 58s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 17m 47s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 2m 40s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 3m 40s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 21m 36s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 15s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 11s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 58s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 7m 0s{color} | | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 27s{color} | | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 38s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 20m 47s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 20m 47s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 18m 18s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 18m 18s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} blanks {color} | {color:green} 0m 0s{color} | | {color:green} The patch has no blanks issues. {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 2m 32s{color} | [/results-checkstyle-root.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/266/artifact/out/results-checkstyle-root.txt] | {color:orange} root: The patch generated 3 new + 413 unchanged - 0 fixed = 416 total (was 413) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 3m 53s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 1s{color} | | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 32s{color} | | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 3m 15s{color} | | {color:green} the patch passed
[jira] [Commented] (YARN-10475) Scale RM-NM heartbeat interval based on node utilization
[ https://issues.apache.org/jira/browse/YARN-10475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222481#comment-17222481 ] Eric Payne commented on YARN-10475: --- [~Jim_Brennan], please add documentation for the new config properties. > Scale RM-NM heartbeat interval based on node utilization > > > Key: YARN-10475 > URL: https://issues.apache.org/jira/browse/YARN-10475 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn >Affects Versions: 2.10.1, 3.4.1 >Reporter: Jim Brennan >Assignee: Jim Brennan >Priority: Minor > Attachments: YARN-10475.001.patch, YARN-10475.002.patch > > > Add the ability to scale the RM-NM heartbeat interval based on node cpu > utilization compared to overall cluster cpu utilization. If a node is > over-utilized compared to the rest of the cluster, it's heartbeat interval > slows down. If it is under-utilized compared to the rest of the cluster, > it's heartbeat interval speeds up. > This is a feature we have been running with internally in production for > several years. It was developed by [~nroberts], based on the observation > that larger faster nodes on our cluster were under-utilized compared to > smaller slower nodes. > This feature is dependent on [YARN-10450], which added cluster-wide > utilization metrics. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222471#comment-17222471 ] Eric Payne commented on YARN-10471: --- Thanks [~Jim_Brennan]. I have uploaded a new patch. > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.004.patch, YARN.10471.005.patch, > YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Payne updated YARN-10471: -- Attachment: YARN.10471.005.patch > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.004.patch, YARN.10471.005.patch, > YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Resolved] (YARN-10477) runc launch failure should not cause nodemanager to go unhealthy
[ https://issues.apache.org/jira/browse/YARN-10477?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jim Brennan resolved YARN-10477. Resolution: Invalid Closing this as invalid. The problem was only there in our internal version of container-executor. I should have checked the code in trunk before filing. > runc launch failure should not cause nodemanager to go unhealthy > > > Key: YARN-10477 > URL: https://issues.apache.org/jira/browse/YARN-10477 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn >Affects Versions: 3.3.1, 3.4.1 >Reporter: Jim Brennan >Assignee: Jim Brennan >Priority: Major > > We have observed some failures when launching containers with runc. We have > not yet identified the root cause of those failures, but a side-effect of > these failures was the Nodemanager marked itself unhealthy. Since these are > rare failures that only affect a single launch, they should not cause the > Nodemanager to be marked unhealthy. > Here is an example RM log: > {noformat} > resourcemanager.log.2020-10-02-03.bz2:2020-10-02 03:20:10,255 [RM Event > dispatcher] INFO rmnode.RMNodeImpl: Node node:8041 reported UNHEALTHY with > details: Linux Container Executor reached unrecoverable exception > {noformat} > And here is an example of the NM log: > {noformat} > 2020-10-02 03:20:02,033 [ContainersLauncher #434] INFO > runtime.RuncContainerRuntime: Launch container failed for > container_e25_1601602719874_10691_01_001723 > org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException: > ExitCodeException exitCode=24: OCI command has bad/missing local dire > ctories > {noformat} > The problem is that the runc code in container-executor is re-using exit code > 24 (INVALID_CONFIG_FILE) which is intended for problems with the > container-executor.cfg file, and those failures are fatal for the NM. We > should use a different exit code for these. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222409#comment-17222409 ] Jim Brennan commented on YARN-10471: Thanks [~epayne]! It looks like there is a problem in the last line in Nodemanager.md. The line appears to be split in two. > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.004.patch, YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222309#comment-17222309 ] Jonathan Hung edited comment on YARN-10467 at 10/28/20, 8:14 PM: - I committed this to trunk/branch-3.3/branch-3.2/branch-3.1/branch-2.10. Thanks [~haibochen] for the contribution and [~Jim_Brennan] for the review. was (Author: jhung): I committed this to trunk~branch-2.10. Thanks [~haibochen] for the contribution and [~Jim_Brennan] for the review. > ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers > - > > Key: YARN-10467 > URL: https://issues.apache.org/jira/browse/YARN-10467 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.10.0, 3.0.3, 3.2.1, 3.1.4 >Reporter: Haibo Chen >Assignee: Haibo Chen >Priority: Major > Fix For: 3.4.0, 3.3.1, 3.1.5, 2.10.2, 3.2.3 > > Attachments: YARN-10467.00.patch, YARN-10467.01.patch, > YARN-10467.02.patch, YARN-10467.branch-2.10.00.patch, > YARN-10467.branch-2.10.01.patch, YARN-10467.branch-2.10.02.patch, > YARN-10467.branch-2.10.03.patch > > > In one of our recent heap analysis, we found that the majority of the heap is > occupied by {{RMNodeImpl.completedContainers}}, which > accounts for 19GB, out of 24.3 GB. There are over 86 million > ContainerIdPBImpl objects, in contrast, only 161,601 RMContainerImpl objects > which represent the # of active containers that RM is still tracking. > Inspecting some ContainerIdPBImpl objects, they belong to applications that > have long finished. This indicates some sort of memory leak of > ContainerIdPBImpl objects in RMNodeImpl. > > Right now, when a container is reported by a NM as completed, it is > immediately added to RMNodeImpl.completedContainers and later cleaned up > after the AM has been notified of its completion in the AM-RM heartbeat. The > cleanup can be broken into a few steps. > * Step 1: the completed container is first added to > RMAppAttemptImpl.justFinishedContainers (this is asynchronous to being added > to {{RMNodeImpl.completedContainers}}). > * Step 2: During the heartbeat AM-RM heartbeat, the container is removed > from RMAppAttemptImpl.justFinishedContainers and added to > RMAppAttemptImpl.finishedContainersSentToAM > Once a completed container gets added to > RMAppAttemptImpl.finishedContainersSentToAM, it is guaranteed to be cleaned > up from {{RMNodeImpl.completedContainers}} > > However, if the AM exits (regardless of failure or success) before some > recently completed containers can be added to > RMAppAttemptImpl.finishedContainersSentToAM in previous heartbeats, there > won’t be any future AM-RM heartbeat to perform aforementioned step 2. Hence, > these objects stay in RMNodeImpl.completedContainers forever. > We have observed in MR that AMs can decide to exit upon success of all it > tasks without waiting for notification of the completion of every container, > or AM may just die suddenly (e.g. OOM). Spark and other framework may just > be similar. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-10477) runc launch failure should not cause nodemanager to go unhealthy
Jim Brennan created YARN-10477: -- Summary: runc launch failure should not cause nodemanager to go unhealthy Key: YARN-10477 URL: https://issues.apache.org/jira/browse/YARN-10477 Project: Hadoop YARN Issue Type: Bug Components: yarn Affects Versions: 3.3.1, 3.4.1 Reporter: Jim Brennan Assignee: Jim Brennan We have observed some failures when launching containers with runc. We have not yet identified the root cause of those failures, but a side-effect of these failures was the Nodemanager marked itself unhealthy. Since these are rare failures that only affect a single launch, they should not cause the Nodemanager to be marked unhealthy. Here is an example RM log: {noformat} resourcemanager.log.2020-10-02-03.bz2:2020-10-02 03:20:10,255 [RM Event dispatcher] INFO rmnode.RMNodeImpl: Node node:8041 reported UNHEALTHY with details: Linux Container Executor reached unrecoverable exception {noformat} And here is an example of the NM log: {noformat} 2020-10-02 03:20:02,033 [ContainersLauncher #434] INFO runtime.RuncContainerRuntime: Launch container failed for container_e25_1601602719874_10691_01_001723 org.apache.hadoop.yarn.server.nodemanager.containermanager.linux.privileged.PrivilegedOperationException: ExitCodeException exitCode=24: OCI command has bad/missing local dire ctories {noformat} The problem is that the runc code in container-executor is re-using exit code 24 (INVALID_CONFIG_FILE) which is intended for problems with the container-executor.cfg file, and those failures are fatal for the NM. We should use a different exit code for these. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222396#comment-17222396 ] Eric Payne commented on YARN-10471: --- Thanks a lot, [~Jim_Brennan], for reviewing these patches. I uploaded version 004. There is no difference from version 003 except that I added some documentation describing the new config properties in NodeManager.md. > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.004.patch, YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Payne updated YARN-10471: -- Attachment: YARN.10471.004.patch > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.004.patch, YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10425) Replace the legacy placement engine in CS with the new one
[ https://issues.apache.org/jira/browse/YARN-10425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222391#comment-17222391 ] Hadoop QA commented on YARN-10425: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 43s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 1s{color} | | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 7 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 49s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 51s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 39s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 57s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 29s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 39s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 39s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 1m 46s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 44s{color} | | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 52s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 57s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 57s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 47s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 47s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} blanks {color} | {color:green} 0m 0s{color} | | {color:green} The patch has no blanks issues. {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 30s{color} | [/results-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/265/artifact/out/results-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt] | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 7 new + 241 unchanged - 8 fixed = 248 total (was 249) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 48s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 22s{color} | | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 39s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 36s{color} | | {color:green} the patch passed with JDK Private
[jira] [Commented] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222385#comment-17222385 ] Hadoop QA commented on YARN-10467: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 21s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 23m 1s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 2s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 52s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 36s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 58s{color} | | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 39s{color} | | {color:green} trunk passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 35s{color} | | {color:green} trunk passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 1m 50s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 46s{color} | | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 52s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 56s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 45s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 45s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} blanks {color} | {color:green} 0m 0s{color} | | {color:green} The patch has no blanks issues. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 33s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 58s{color} | | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 39s{color} | | {color:green} the patch passed with JDK Ubuntu-11.0.8+10-post-Ubuntu-0ubuntu118.04.1 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 35s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~18.04-b01 {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 59s{color} | | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 92m 59s{color} | | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color}
[jira] [Commented] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222318#comment-17222318 ] Hadoop QA commented on YARN-10467: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 23s{color} | | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} branch-2.10 Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 5s{color} | | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 47s{color} | | {color:green} branch-2.10 passed with JDK Oracle Corporation-1.7.0_95-b00 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 40s{color} | | {color:green} branch-2.10 passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~16.04-b01 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 27s{color} | | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 45s{color} | | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 33s{color} | | {color:green} branch-2.10 passed with JDK Oracle Corporation-1.7.0_95-b00 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 27s{color} | | {color:green} branch-2.10 passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~16.04-b01 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 1m 31s{color} | | {color:blue} Used deprecated FindBugs config; considering switching to SpotBugs. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 29s{color} | | {color:green} branch-2.10 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 40s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 43s{color} | | {color:green} the patch passed with JDK Oracle Corporation-1.7.0_95-b00 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 43s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 37s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~16.04-b01 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 37s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} blanks {color} | {color:green} 0m 0s{color} | | {color:green} The patch has no blanks issues. {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 47s{color} | | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 31s{color} | | {color:green} the patch passed with JDK Oracle Corporation-1.7.0_95-b00 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 26s{color} | | {color:green} the patch passed with JDK Private Build-1.8.0_265-8u265-b01-0ubuntu2~16.04-b01 {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 57s{color} | | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 68m 48s{color} | [/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt|https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/263/artifact/out/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt] | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 24s{color} | | {color:green} The patch does not generate ASF
[jira] [Commented] (YARN-10475) Scale RM-NM heartbeat interval based on node utilization
[ https://issues.apache.org/jira/browse/YARN-10475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1794#comment-1794 ] Jim Brennan commented on YARN-10475: I put up patch 002 to address checkstyle/javac issues. > Scale RM-NM heartbeat interval based on node utilization > > > Key: YARN-10475 > URL: https://issues.apache.org/jira/browse/YARN-10475 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn >Affects Versions: 2.10.1, 3.4.1 >Reporter: Jim Brennan >Assignee: Jim Brennan >Priority: Minor > Attachments: YARN-10475.001.patch, YARN-10475.002.patch > > > Add the ability to scale the RM-NM heartbeat interval based on node cpu > utilization compared to overall cluster cpu utilization. If a node is > over-utilized compared to the rest of the cluster, it's heartbeat interval > slows down. If it is under-utilized compared to the rest of the cluster, > it's heartbeat interval speeds up. > This is a feature we have been running with internally in production for > several years. It was developed by [~nroberts], based on the observation > that larger faster nodes on our cluster were under-utilized compared to > smaller slower nodes. > This feature is dependent on [YARN-10450], which added cluster-wide > utilization metrics. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10475) Scale RM-NM heartbeat interval based on node utilization
[ https://issues.apache.org/jira/browse/YARN-10475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jim Brennan updated YARN-10475: --- Attachment: YARN-10475.002.patch > Scale RM-NM heartbeat interval based on node utilization > > > Key: YARN-10475 > URL: https://issues.apache.org/jira/browse/YARN-10475 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn >Affects Versions: 2.10.1, 3.4.1 >Reporter: Jim Brennan >Assignee: Jim Brennan >Priority: Minor > Attachments: YARN-10475.001.patch, YARN-10475.002.patch > > > Add the ability to scale the RM-NM heartbeat interval based on node cpu > utilization compared to overall cluster cpu utilization. If a node is > over-utilized compared to the rest of the cluster, it's heartbeat interval > slows down. If it is under-utilized compared to the rest of the cluster, > it's heartbeat interval speeds up. > This is a feature we have been running with internally in production for > several years. It was developed by [~nroberts], based on the observation > that larger faster nodes on our cluster were under-utilized compared to > smaller slower nodes. > This feature is dependent on [YARN-10450], which added cluster-wide > utilization metrics. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10425) Replace the legacy placement engine in CS with the new one
[ https://issues.apache.org/jira/browse/YARN-10425?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1777#comment-1777 ] Gergely Pollak commented on YARN-10425: --- Patch#3 is a rebased version and addressing a major issue during recovery, but I need to see if it broke anything else. Review comment related fixes are expected when I've dealt with the bug. [~wangda] thank you, yes backwards compatibility is of utmost importance, there might be some slight differences in the error handling, due to the inconsistencies of the legacy solution, but we also provide a way to define on a per rule basis what should happen in the case of an error (error when the placement rule cannot be executed eg. invalid target path). > Replace the legacy placement engine in CS with the new one > -- > > Key: YARN-10425 > URL: https://issues.apache.org/jira/browse/YARN-10425 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Gergely Pollak >Assignee: Gergely Pollak >Priority: Major > Attachments: YARN-10425.001.patch, YARN-10425.002.patch, > YARN-10425.003.patch > > > Remove the UserGroupMapping and ApplicationName mapping classes, and use the > new CSMappingPlacementRule instead. Also cleanup the orphan classes which are > used by these classes only. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10425) Replace the legacy placement engine in CS with the new one
[ https://issues.apache.org/jira/browse/YARN-10425?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gergely Pollak updated YARN-10425: -- Attachment: YARN-10425.003.patch > Replace the legacy placement engine in CS with the new one > -- > > Key: YARN-10425 > URL: https://issues.apache.org/jira/browse/YARN-10425 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Gergely Pollak >Assignee: Gergely Pollak >Priority: Major > Attachments: YARN-10425.001.patch, YARN-10425.002.patch, > YARN-10425.003.patch > > > Remove the UserGroupMapping and ApplicationName mapping classes, and use the > new CSMappingPlacementRule instead. Also cleanup the orphan classes which are > used by these classes only. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Haibo Chen updated YARN-10467: -- Attachment: YARN-10467.02.patch > ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers > - > > Key: YARN-10467 > URL: https://issues.apache.org/jira/browse/YARN-10467 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.10.0, 3.0.3, 3.2.1, 3.1.4 >Reporter: Haibo Chen >Assignee: Haibo Chen >Priority: Major > Attachments: YARN-10467.00.patch, YARN-10467.01.patch, > YARN-10467.02.patch, YARN-10467.branch-2.10.00.patch, > YARN-10467.branch-2.10.01.patch, YARN-10467.branch-2.10.02.patch, > YARN-10467.branch-2.10.03.patch > > > In one of our recent heap analysis, we found that the majority of the heap is > occupied by {{RMNodeImpl.completedContainers}}, which > accounts for 19GB, out of 24.3 GB. There are over 86 million > ContainerIdPBImpl objects, in contrast, only 161,601 RMContainerImpl objects > which represent the # of active containers that RM is still tracking. > Inspecting some ContainerIdPBImpl objects, they belong to applications that > have long finished. This indicates some sort of memory leak of > ContainerIdPBImpl objects in RMNodeImpl. > > Right now, when a container is reported by a NM as completed, it is > immediately added to RMNodeImpl.completedContainers and later cleaned up > after the AM has been notified of its completion in the AM-RM heartbeat. The > cleanup can be broken into a few steps. > * Step 1: the completed container is first added to > RMAppAttemptImpl.justFinishedContainers (this is asynchronous to being added > to {{RMNodeImpl.completedContainers}}). > * Step 2: During the heartbeat AM-RM heartbeat, the container is removed > from RMAppAttemptImpl.justFinishedContainers and added to > RMAppAttemptImpl.finishedContainersSentToAM > Once a completed container gets added to > RMAppAttemptImpl.finishedContainersSentToAM, it is guaranteed to be cleaned > up from {{RMNodeImpl.completedContainers}} > > However, if the AM exits (regardless of failure or success) before some > recently completed containers can be added to > RMAppAttemptImpl.finishedContainersSentToAM in previous heartbeats, there > won’t be any future AM-RM heartbeat to perform aforementioned step 2. Hence, > these objects stay in RMNodeImpl.completedContainers forever. > We have observed in MR that AMs can decide to exit upon success of all it > tasks without waiting for notification of the completion of every container, > or AM may just die suddenly (e.g. OOM). Spark and other framework may just > be similar. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Haibo Chen updated YARN-10467: -- Attachment: YARN-10467.branch-2.10.03.patch > ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers > - > > Key: YARN-10467 > URL: https://issues.apache.org/jira/browse/YARN-10467 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.10.0, 3.0.3, 3.2.1, 3.1.4 >Reporter: Haibo Chen >Assignee: Haibo Chen >Priority: Major > Attachments: YARN-10467.00.patch, YARN-10467.01.patch, > YARN-10467.branch-2.10.00.patch, YARN-10467.branch-2.10.01.patch, > YARN-10467.branch-2.10.02.patch, YARN-10467.branch-2.10.03.patch > > > In one of our recent heap analysis, we found that the majority of the heap is > occupied by {{RMNodeImpl.completedContainers}}, which > accounts for 19GB, out of 24.3 GB. There are over 86 million > ContainerIdPBImpl objects, in contrast, only 161,601 RMContainerImpl objects > which represent the # of active containers that RM is still tracking. > Inspecting some ContainerIdPBImpl objects, they belong to applications that > have long finished. This indicates some sort of memory leak of > ContainerIdPBImpl objects in RMNodeImpl. > > Right now, when a container is reported by a NM as completed, it is > immediately added to RMNodeImpl.completedContainers and later cleaned up > after the AM has been notified of its completion in the AM-RM heartbeat. The > cleanup can be broken into a few steps. > * Step 1: the completed container is first added to > RMAppAttemptImpl.justFinishedContainers (this is asynchronous to being added > to {{RMNodeImpl.completedContainers}}). > * Step 2: During the heartbeat AM-RM heartbeat, the container is removed > from RMAppAttemptImpl.justFinishedContainers and added to > RMAppAttemptImpl.finishedContainersSentToAM > Once a completed container gets added to > RMAppAttemptImpl.finishedContainersSentToAM, it is guaranteed to be cleaned > up from {{RMNodeImpl.completedContainers}} > > However, if the AM exits (regardless of failure or success) before some > recently completed containers can be added to > RMAppAttemptImpl.finishedContainersSentToAM in previous heartbeats, there > won’t be any future AM-RM heartbeat to perform aforementioned step 2. Hence, > these objects stay in RMNodeImpl.completedContainers forever. > We have observed in MR that AMs can decide to exit upon success of all it > tasks without waiting for notification of the completion of every container, > or AM may just die suddenly (e.g. OOM). Spark and other framework may just > be similar. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1738#comment-1738 ] Haibo Chen commented on YARN-10467: --- Thanks for catching this, [~Jim_Brennan]! I'll quickly update the patch to address this. > ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers > - > > Key: YARN-10467 > URL: https://issues.apache.org/jira/browse/YARN-10467 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.10.0, 3.0.3, 3.2.1, 3.1.4 >Reporter: Haibo Chen >Assignee: Haibo Chen >Priority: Major > Attachments: YARN-10467.00.patch, YARN-10467.01.patch, > YARN-10467.branch-2.10.00.patch, YARN-10467.branch-2.10.01.patch, > YARN-10467.branch-2.10.02.patch > > > In one of our recent heap analysis, we found that the majority of the heap is > occupied by {{RMNodeImpl.completedContainers}}, which > accounts for 19GB, out of 24.3 GB. There are over 86 million > ContainerIdPBImpl objects, in contrast, only 161,601 RMContainerImpl objects > which represent the # of active containers that RM is still tracking. > Inspecting some ContainerIdPBImpl objects, they belong to applications that > have long finished. This indicates some sort of memory leak of > ContainerIdPBImpl objects in RMNodeImpl. > > Right now, when a container is reported by a NM as completed, it is > immediately added to RMNodeImpl.completedContainers and later cleaned up > after the AM has been notified of its completion in the AM-RM heartbeat. The > cleanup can be broken into a few steps. > * Step 1: the completed container is first added to > RMAppAttemptImpl.justFinishedContainers (this is asynchronous to being added > to {{RMNodeImpl.completedContainers}}). > * Step 2: During the heartbeat AM-RM heartbeat, the container is removed > from RMAppAttemptImpl.justFinishedContainers and added to > RMAppAttemptImpl.finishedContainersSentToAM > Once a completed container gets added to > RMAppAttemptImpl.finishedContainersSentToAM, it is guaranteed to be cleaned > up from {{RMNodeImpl.completedContainers}} > > However, if the AM exits (regardless of failure or success) before some > recently completed containers can be added to > RMAppAttemptImpl.finishedContainersSentToAM in previous heartbeats, there > won’t be any future AM-RM heartbeat to perform aforementioned step 2. Hence, > these objects stay in RMNodeImpl.completedContainers forever. > We have observed in MR that AMs can decide to exit upon success of all it > tasks without waiting for notification of the completion of every container, > or AM may just die suddenly (e.g. OOM). Spark and other framework may just > be similar. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10467) ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers
[ https://issues.apache.org/jira/browse/YARN-10467?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1735#comment-1735 ] Jim Brennan commented on YARN-10467: Thanks for reporting this and for the solution [~haibochen]! Everything looks good to me. I hesitate to mention one minor nit, a typo in this comment: {quote}// there might be some completed containers that *are have* not been pulled {quote} It's up to you whether you want to fix this. [~jhung] were you planning to commit this? > ContainerIdPBImpl objects can be leaked in RMNodeImpl.completedContainers > - > > Key: YARN-10467 > URL: https://issues.apache.org/jira/browse/YARN-10467 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.10.0, 3.0.3, 3.2.1, 3.1.4 >Reporter: Haibo Chen >Assignee: Haibo Chen >Priority: Major > Attachments: YARN-10467.00.patch, YARN-10467.01.patch, > YARN-10467.branch-2.10.00.patch, YARN-10467.branch-2.10.01.patch, > YARN-10467.branch-2.10.02.patch > > > In one of our recent heap analysis, we found that the majority of the heap is > occupied by {{RMNodeImpl.completedContainers}}, which > accounts for 19GB, out of 24.3 GB. There are over 86 million > ContainerIdPBImpl objects, in contrast, only 161,601 RMContainerImpl objects > which represent the # of active containers that RM is still tracking. > Inspecting some ContainerIdPBImpl objects, they belong to applications that > have long finished. This indicates some sort of memory leak of > ContainerIdPBImpl objects in RMNodeImpl. > > Right now, when a container is reported by a NM as completed, it is > immediately added to RMNodeImpl.completedContainers and later cleaned up > after the AM has been notified of its completion in the AM-RM heartbeat. The > cleanup can be broken into a few steps. > * Step 1: the completed container is first added to > RMAppAttemptImpl.justFinishedContainers (this is asynchronous to being added > to {{RMNodeImpl.completedContainers}}). > * Step 2: During the heartbeat AM-RM heartbeat, the container is removed > from RMAppAttemptImpl.justFinishedContainers and added to > RMAppAttemptImpl.finishedContainersSentToAM > Once a completed container gets added to > RMAppAttemptImpl.finishedContainersSentToAM, it is guaranteed to be cleaned > up from {{RMNodeImpl.completedContainers}} > > However, if the AM exits (regardless of failure or success) before some > recently completed containers can be added to > RMAppAttemptImpl.finishedContainersSentToAM in previous heartbeats, there > won’t be any future AM-RM heartbeat to perform aforementioned step 2. Hence, > these objects stay in RMNodeImpl.completedContainers forever. > We have observed in MR that AMs can decide to exit upon success of all it > tasks without waiting for notification of the completion of every container, > or AM may just die suddenly (e.g. OOM). Spark and other framework may just > be similar. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1723#comment-1723 ] Jim Brennan commented on YARN-10471: Thanks for putting this up [~epayne]! I am +1 on patches for trunk and branch-3.2. I will wait until tomorrow to commit this, to give others a chance to chime in if desired. > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10471) Prevent logs for any container from becoming larger than a configurable size.
[ https://issues.apache.org/jira/browse/YARN-10471?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=1715#comment-1715 ] Eric Payne commented on YARN-10471: --- The unit tests that failed in the branch 2 pre-commit build also fail without this patch in the same way. > Prevent logs for any container from becoming larger than a configurable size. > - > > Key: YARN-10471 > URL: https://issues.apache.org/jira/browse/YARN-10471 > Project: Hadoop YARN > Issue Type: Improvement >Affects Versions: 3.2.1, 3.1.4 >Reporter: Eric Payne >Assignee: Eric Payne >Priority: Minor > Attachments: YARN.10471.001.patch, YARN.10471.002.patch, > YARN.10471.003.patch, YARN.10471.branch-3.2.003.patch > > > Configure a cluster such that a task attempt will be killed if any container > log exceeds a configured size. This would help prevent logs from filling > disks and also prevent the need to aggregate enormous logs. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10458) Hive On Tez queries fails upon submission to dynamically created pools
[ https://issues.apache.org/jira/browse/YARN-10458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17222114#comment-17222114 ] Peter Bacsko commented on YARN-10458: - [~leftnoteasy] I'll add another test case for that. > Hive On Tez queries fails upon submission to dynamically created pools > -- > > Key: YARN-10458 > URL: https://issues.apache.org/jira/browse/YARN-10458 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Reporter: Anand Srinivasan >Assignee: Peter Bacsko >Priority: Major > Attachments: YARN-10458-001.patch, YARN-10458-002.patch > > > While using Dynamic Auto-Creation and Management of Leaf Queues, we could see > that the queue creation fails because ACL submit application check couldn't > succeed. > We tried setting acl_submit_applications to '*' for managed parent queues. > For static queues, this worked but failed for dynamic queues. Also tried > setting the below property but it didn't help either. > yarn.scheduler.capacity.root.parent-queue-name.leaf-queue-template.acl_submit_applications=*. > RM error log shows the following : > 2020-09-18 01:08:40,579 INFO > org.apache.hadoop.yarn.server.resourcemanager.placement.UserGroupMappingPlacementRule: > Application application_1600399068816_0460 user user1 mapping [default] to > [queue1] override false > 2020-09-18 01:08:40,579 WARN > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager: User 'user1' from > application tag does not have access to queue 'user1'. The placement is done > for user 'hive' > > Checking the code, scheduler#checkAccess() bails out even before checking the > ACL permissions for that particular queue because the CSQueue is null. > {code:java} > public boolean checkAccess(UserGroupInformation callerUGI, > QueueACL acl, String queueName) { > CSQueue queue = getQueue(queueName); > if (queue == null) { > if (LOG.isDebugEnabled()) > { LOG.debug("ACL not found for queue access-type " + acl + " for queue " + > queueName); } > return false;*<-- the method returns false here.* > } > return queue.hasAccess(acl, callerUGI); > } > {code} > As this is an auto created queue, CSQueue may be null in this case. May be > scheduler#checkAccess() should have a logic to differentiate when CSQueue is > null and if queue mapping is involved and if so, check if the parent queue > exists and is a managed parent and if so, check if the parent queue has valid > ACL's instead of returning false ? > Thanks -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10476) Queue metrics of Unmanaged applications
[ https://issues.apache.org/jira/browse/YARN-10476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10476: - Description: Right now we do not have separate metrics unmanaged applications. All application metrics come as part of Queue (Managed and UnManaged), This Jira aims to show them separately. > Queue metrics of Unmanaged applications > > > Key: YARN-10476 > URL: https://issues.apache.org/jira/browse/YARN-10476 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Minor > > Right now we do not have separate metrics unmanaged applications. All > application metrics come as part of Queue (Managed and UnManaged), This Jira > aims to show them separately. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-10476) Queue metrics of Unmanaged applications
Cyrus Jackson created YARN-10476: Summary: Queue metrics of Unmanaged applications Key: YARN-10476 URL: https://issues.apache.org/jira/browse/YARN-10476 Project: Hadoop YARN Issue Type: Bug Components: resourcemanager Reporter: Cyrus Jackson Assignee: Cyrus Jackson -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org