[jira] [Commented] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384697#comment-17384697 ] Andras Gyori commented on YARN-10860: - Thanks [~ebadger] for the patch. It is straightforward. +1 non-binding. > Make max container per heartbeat configs refreshable > > > Key: YARN-10860 > URL: https://issues.apache.org/jira/browse/YARN-10860 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Eric Badger >Assignee: Eric Badger >Priority: Major > Attachments: YARN-10860.001.patch, YARN-10860.branch-2.10.001.patch > > > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-container-assignments}} > and > {{yarn.scheduler.capacity.per-node-heartbeat.multiple-assignments-enabled}} > are currently *not* refreshable configs, but I believe they should be. This > JIRA is to turn these into refreshable configs, just like > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments}} > is. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10867) YARN should expose a ENV used to map a custom device into docker container
[ https://issues.apache.org/jira/browse/YARN-10867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384623#comment-17384623 ] Chi Heng commented on YARN-10867: - Hi [~ebadger] First,regardless of security issues,{{YARN_CONTAINER_RUNTIME_DOCKER_MOUNTS will be turned to an -v arg when docker container starts(docker do this using Linux binding mount).I suppose it is not suitable for /dev/fuse,GPU,RDMA or other devices.Actually I tried to mount device as a volume into docker,but the device cant't be used in container.}} And back to security issues, actually I just need some docker capabilites like 'SYS_ADMIN' .But since I can't mount a device into container,I had to request a privileged docker container to complete this job ('–privileged'),where devices is inluded by default.Thus the problem became complex.The privileged container runs with root user,so the log aggregation can't work normally.And the user which submitted the application is used to call the docker command, so I had to add all non-privileged user into docker group . I suppose it is not a good idea using privileged container to complete this job, expose an new ENV like '{{YARN_CONTAINER_RUNTIME_DOCKER_DEVICES}}' may be a better option > YARN should expose a ENV used to map a custom device into docker container > -- > > Key: YARN-10867 > URL: https://issues.apache.org/jira/browse/YARN-10867 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Chi Heng >Priority: Major > > In some scenarios, like mounting a FUSE in docker,user needs to map a custom > device (eg. /dev/fuse) into docker container.I notice that an adddevice > method is defined in [ > hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java > > |https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] > ,I suppose that an ENV or config property should to be exposed to user to > call this method -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384603#comment-17384603 ] Qi Zhu commented on YARN-10860: --- LGTM , if [~gandras] other comments. > Make max container per heartbeat configs refreshable > > > Key: YARN-10860 > URL: https://issues.apache.org/jira/browse/YARN-10860 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Eric Badger >Assignee: Eric Badger >Priority: Major > Attachments: YARN-10860.001.patch, YARN-10860.branch-2.10.001.patch > > > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-container-assignments}} > and > {{yarn.scheduler.capacity.per-node-heartbeat.multiple-assignments-enabled}} > are currently *not* refreshable configs, but I believe they should be. This > JIRA is to turn these into refreshable configs, just like > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments}} > is. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10657) We should make max application per queue to support node label.
[ https://issues.apache.org/jira/browse/YARN-10657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384594#comment-17384594 ] Qi Zhu commented on YARN-10657: --- Thanks [~gandras] for update, LGTM now, just fix the checkstyle. > We should make max application per queue to support node label. > --- > > Key: YARN-10657 > URL: https://issues.apache.org/jira/browse/YARN-10657 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Qi Zhu >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10657.001.patch, YARN-10657.002.patch, > YARN-10657.003.patch, YARN-10657.004.patch > > > https://issues.apache.org/jira/browse/YARN-10641?focusedCommentId=17291708&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17291708 > As we discussed in above comment: > We should deep into the label related max applications per queue. > I think when node label enabled in queue, max applications should consider > the max capacity of all labels. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10630) [UI2] Ambiguous queue name resolution
[ https://issues.apache.org/jira/browse/YARN-10630?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384573#comment-17384573 ] Qi Zhu commented on YARN-10630: --- Committed to trunk, thanks [~gandras] ! > [UI2] Ambiguous queue name resolution > - > > Key: YARN-10630 > URL: https://issues.apache.org/jira/browse/YARN-10630 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn-ui-v2 >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > Labels: pull-request-available > Attachments: Screenshot 2021-07-19 at 15.30.38.png > > Time Spent: 1h 20m > Remaining Estimate: 0h > > Yarn UIv2 uses queueName instead of queuePath (which was added in the > scheduler response in YARN-10610), which makes the queue resolution ambiguous > in case of identical queue short names (eg. root.a.b <-> root.b). This causes > invalid behaviour in multiple places. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Resolved] (YARN-10630) [UI2] Ambiguous queue name resolution
[ https://issues.apache.org/jira/browse/YARN-10630?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Qi Zhu resolved YARN-10630. --- Fix Version/s: 3.4.0 Resolution: Fixed > [UI2] Ambiguous queue name resolution > - > > Key: YARN-10630 > URL: https://issues.apache.org/jira/browse/YARN-10630 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn-ui-v2 >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > Labels: pull-request-available > Fix For: 3.4.0 > > Attachments: Screenshot 2021-07-19 at 15.30.38.png > > Time Spent: 1h 20m > Remaining Estimate: 0h > > Yarn UIv2 uses queueName instead of queuePath (which was added in the > scheduler response in YARN-10610), which makes the queue resolution ambiguous > in case of identical queue short names (eg. root.a.b <-> root.b). This causes > invalid behaviour in multiple places. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384560#comment-17384560 ] Hadoop QA commented on YARN-10860: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 42s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} markdownlint {color} | {color:blue} 0m 1s{color} | {color:blue}{color} | {color:blue} markdownlint was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red}{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 44s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 28s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 21s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 9s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 41s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 54s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 8s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 37s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 35s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 23m 0s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 38s{color} | {color:blue}{color} | {color:blue} branch/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site no spotbugs output file (spotbugsXml.xml) {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 27s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 7s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 46s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 46s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 10s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 10s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 37s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 44s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:gre
[jira] [Commented] (YARN-10867) YARN should expose a ENV used to map a custom device into docker container
[ https://issues.apache.org/jira/browse/YARN-10867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384513#comment-17384513 ] Eric Badger commented on YARN-10867: https://hadoop.apache.org/docs/stable/hadoop-yarn/hadoop-yarn-site/DockerContainers.html I believe you can just use {{docker.allowed.devices}} in your container-executor.cfg file if you need to mount an actual device. However, you'll need to be a privileged container to do that, so you'll need to also set {{docker.privileged-containers.enabled=true}}. Note that running privileged containers is very risky and adds a lot of security concerns with it, so proceed with caution. After setting those, I believe you can use {{YARN_CONTAINER_RUNTIME_DOCKER_MOUNTS}} to specify the mounts that you want, including the device such as {{/dev/fuse}} > YARN should expose a ENV used to map a custom device into docker container > -- > > Key: YARN-10867 > URL: https://issues.apache.org/jira/browse/YARN-10867 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Chi Heng >Priority: Major > > In some scenarios, like mounting a FUSE in docker,user needs to map a custom > device (eg. /dev/fuse) into docker container.I notice that an adddevice > method is defined in [ > hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java > > |https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] > ,I suppose that an ENV or config property should to be exposed to user to > call this method -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Badger updated YARN-10860: --- Attachment: (was: YARN-10860.001.patch) > Make max container per heartbeat configs refreshable > > > Key: YARN-10860 > URL: https://issues.apache.org/jira/browse/YARN-10860 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Eric Badger >Assignee: Eric Badger >Priority: Major > Attachments: YARN-10860.001.patch, YARN-10860.branch-2.10.001.patch > > > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-container-assignments}} > and > {{yarn.scheduler.capacity.per-node-heartbeat.multiple-assignments-enabled}} > are currently *not* refreshable configs, but I believe they should be. This > JIRA is to turn these into refreshable configs, just like > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments}} > is. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Badger updated YARN-10860: --- Attachment: YARN-10860.001.patch > Make max container per heartbeat configs refreshable > > > Key: YARN-10860 > URL: https://issues.apache.org/jira/browse/YARN-10860 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Eric Badger >Assignee: Eric Badger >Priority: Major > Attachments: YARN-10860.001.patch, YARN-10860.branch-2.10.001.patch > > > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-container-assignments}} > and > {{yarn.scheduler.capacity.per-node-heartbeat.multiple-assignments-enabled}} > are currently *not* refreshable configs, but I believe they should be. This > JIRA is to turn these into refreshable configs, just like > {{yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments}} > is. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10860) Make max container per heartbeat configs refreshable
[ https://issues.apache.org/jira/browse/YARN-10860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384499#comment-17384499 ] Hadoop QA commented on YARN-10860: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 9m 38s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} markdownlint {color} | {color:blue} 0m 0s{color} | {color:blue}{color} | {color:blue} markdownlint was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red}{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} branch-2.10 Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 2m 21s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 12m 21s{color} | {color:green}{color} | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 10s{color} | {color:green}{color} | {color:green} branch-2.10 passed with JDK Azul Systems, Inc.-1.7.0_262-b10 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 8s{color} | {color:green}{color} | {color:green} branch-2.10 passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~16.04.1-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 0s{color} | {color:green}{color} | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 34s{color} | {color:green}{color} | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 16s{color} | {color:green}{color} | {color:green} branch-2.10 passed with JDK Azul Systems, Inc.-1.7.0_262-b10 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 9s{color} | {color:green}{color} | {color:green} branch-2.10 passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~16.04.1-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 7m 43s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 34s{color} | {color:blue}{color} | {color:blue} branch/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site no spotbugs output file (spotbugsXml.xml) {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 20s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 54s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 36s{color} | {color:green}{color} | {color:green} the patch passed with JDK Azul Systems, Inc.-1.7.0_262-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 36s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 3s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~16.04.1-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 3s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 54s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 25s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch has no whitespace issues. {color} | | {color:gr
[jira] [Commented] (YARN-10854) Support marking inactive node as untracked without configured include path
[ https://issues.apache.org/jira/browse/YARN-10854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384440#comment-17384440 ] Daniel Templeton commented on YARN-10854: - In principle the change looks fine to me. I'd like the docs in the XML file to be a bit more explicit about what this property does. If you know about the behavior of the include and exclude lists, the explanation is just barely enough to know what the flag does. It would be better to be more verbose, provide some context, make it easier to understand without reading source code. > Support marking inactive node as untracked without configured include path > -- > > Key: YARN-10854 > URL: https://issues.apache.org/jira/browse/YARN-10854 > Project: Hadoop YARN > Issue Type: Improvement > Components: resourcemanager >Reporter: Tao Yang >Assignee: Tao Yang >Priority: Major > Attachments: YARN-10854.001.patch > > > Currently inactive nodes which have been decommissioned/shutdown/lost for a > while(specified expiration time defined via > {{yarn.resourcemanager.node-removal-untracked.timeout-ms}}, 60 seconds by > default) and not exist in both include and exclude files can be marked as > untracked nodes and can be removed from RM state (YARN-4311). It's very > useful when auto-scaling is enabled in elastic cloud environment, which can > avoid unlimited increase of inactive nodes (mostly are decommissioned nodes). > But this only works when the include path is configured, mismatched for most > of our cloud environments without configured white list of nodes, which can > lead to easily control for the auto-scaling of nodes without further security > requirements. > So I propose to support marking inactive node as untracked without configured > include path, to be compatible with the former versions, we can add a switch > config for this. > Any thoughts/suggestions/feedbacks are welcome! -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-10857) YarnClient Caching Addresses
[ https://issues.apache.org/jira/browse/YARN-10857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384426#comment-17384426 ] Steve Suh edited comment on YARN-10857 at 7/20/21, 6:08 PM: InetSocketAddress address objects are created at the initialization of the YarnClient using the hosts defined for the rm’s (in this case rm1 and rm2). If entries for the hosts do not exist in /etc/hosts (or resolvable by dns) during this initialization, then an unresolvable InetSocketAddress object will be created for them and will be passed to the Connection to use. The InetSocketAddress object is passed to and reused by the IPC [Client.java|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java] As can be seen here, the InetSocketAddress is only checked if it's an unresolved address. It does not attempt to re-resolve the cached address and throws. [https://github.com/apache/hadoop/blob/de41ce8a16434aee13f705a9e3666f29e8ec8cb3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java#L1602-L1608] {code:java} if (address.isUnresolved()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} A possible fix would be to change this check and include the following: _if (address.isUnresolved() *{color:#4c9aff}&& !updateAddress(){color}*)_ {code:java} if (address.isUnresolved() && !updateAddress()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} was (Author: suhsteve): InetSocketAddress address objects are created at the initialization of the YarnClient using the hosts defined for the rm’s (in this case rm1 and rm2). If entries for the hosts do not exist in /etc/hosts (or resolvable by dns) during this initialization, then an unresolvable InetSocketAddress object will be created for them and will be passed to the Connection to use. The InetSocketAddress object is passed to and reused by the IPC [Client.java|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java] As can be seen here, the InetSocketAddress is only checked if it's an unresolved address. It does not attempt to re-resolve the cached address and throws and error. [https://github.com/apache/hadoop/blob/de41ce8a16434aee13f705a9e3666f29e8ec8cb3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java#L1602-L1608] {code:java} if (address.isUnresolved()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} A possible fix would be to change this check and include the following: _if (address.isUnresolved() *{color:#4c9aff}&& !updateAddress(){color}*)_ {code:java} if (address.isUnresolved() && !updateAddress()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} > YarnClient Caching Addresses > > > Key: YARN-10857 > URL: https://issues.apache.org/jira/browse/YARN-10857 > Project: Hadoop YARN > Issue Type: Improvement > Components: client, yarn >Reporter: Steve Suh >Assignee: Prabhu Joseph >Priority: Minor > > We have noticed that when the YarnClient is initialized and used, it is not > very resilient when dns or /etc/hosts is modified in the following scenario: > Take for instance the following (and reproducable) sequence of events that > can occur on a service that instantiates and uses YarnClient. > - Yarn has rm HA enabled (*yarn.resourcemanager.ha.enabled* is *true*) and > there are two rms (rm1 and rm2). > - *yarn.client.failover-proxy-provider* is set to > *org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider* > 1)rm2 is currently the active rm > 2)/etc/hosts (or dns) is missing host information for rm2 > 3)A service is started and it initializes the YarnClient at startup. > 4)At some point in time after YarnClient is done initializing, /etc/hosts > is updated and contains host information for rm2 > 5)Yarn is queried, for instance calling *yarnclient.getApplications()* > 6)All YarnClient attempts to communicate with rm2 fail with > UnknownHostExceptions, even though /etc/hosts now contains host information > for it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-iss
[jira] [Updated] (YARN-10355) Refactor NM ContainerLaunch.java#orderEnvByDependencies
[ https://issues.apache.org/jira/browse/YARN-10355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated YARN-10355: -- Labels: pull-request-available (was: ) > Refactor NM ContainerLaunch.java#orderEnvByDependencies > --- > > Key: YARN-10355 > URL: https://issues.apache.org/jira/browse/YARN-10355 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn >Reporter: Benjamin Teke >Assignee: Tamas Domok >Priority: Minor > Labels: pull-request-available > Time Spent: 10m > Remaining Estimate: 0h > > The > {{org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch#orderEnvByDependencies}} > and it's helper method \{{getEnvDependencies }}(together with the overrides) > is hard to read. Some improvements could be made: > * use Pattern matching in the overrides of getEnvDependencies instead of > iterating through the environmental variable strings char by char > * the unit tests contains a lot of repeated code and generally the test > methods are long - they could be separated into different setup/helper and > assertion methods -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-10857) YarnClient Caching Addresses
[ https://issues.apache.org/jira/browse/YARN-10857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384426#comment-17384426 ] Steve Suh edited comment on YARN-10857 at 7/20/21, 5:55 PM: InetSocketAddress address objects are created at the initialization of the YarnClient using the hosts defined for the rm’s (in this case rm1 and rm2). If entries for the hosts do not exist in /etc/hosts (or resolvable by dns) during this initialization, then an unresolvable InetSocketAddress object will be created for them and will be passed to the Connection to use. The InetSocketAddress object is passed to and reused by the IPC [Client.java|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java] As can be seen here, the InetSocketAddress is only checked if it's an unresolved address. It does not attempt to re-resolve the cached address and throws and error. [https://github.com/apache/hadoop/blob/de41ce8a16434aee13f705a9e3666f29e8ec8cb3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java#L1602-L1608] {code:java} if (address.isUnresolved()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} A possible fix would be to change this check and include the following: _if (address.isUnresolved() *{color:#4c9aff}&& !updateAddress(){color}*)_ {code:java} if (address.isUnresolved() && !updateAddress()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} was (Author: suhsteve): An unresolved InetSocketAddress object is created during the initialization of the YarnClient. This InetSocketAddress object is passed around and reused by the IPC [Client.java|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java] Here the InetSocketAddress is only checked if it's an unresolved address. It does not attempt to re-resolve the cached address and throws and error. https://github.com/apache/hadoop/blob/de41ce8a16434aee13f705a9e3666f29e8ec8cb3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java#L1602-L1608 {code:java} if (address.isUnresolved()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} A possible fix would be to change this check and include the following: _if (address.isUnresolved() *{color:#4C9AFF}&& !updateAddress(){color}*)_ {code:java} if (address.isUnresolved() && !updateAddress()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} > YarnClient Caching Addresses > > > Key: YARN-10857 > URL: https://issues.apache.org/jira/browse/YARN-10857 > Project: Hadoop YARN > Issue Type: Improvement > Components: client, yarn >Reporter: Steve Suh >Assignee: Prabhu Joseph >Priority: Minor > > We have noticed that when the YarnClient is initialized and used, it is not > very resilient when dns or /etc/hosts is modified in the following scenario: > Take for instance the following (and reproducable) sequence of events that > can occur on a service that instantiates and uses YarnClient. > - Yarn has rm HA enabled (*yarn.resourcemanager.ha.enabled* is *true*) and > there are two rms (rm1 and rm2). > - *yarn.client.failover-proxy-provider* is set to > *org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider* > 1)rm2 is currently the active rm > 2)/etc/hosts (or dns) is missing host information for rm2 > 3)A service is started and it initializes the YarnClient at startup. > 4)At some point in time after YarnClient is done initializing, /etc/hosts > is updated and contains host information for rm2 > 5)Yarn is queried, for instance calling *yarnclient.getApplications()* > 6)All YarnClient attempts to communicate with rm2 fail with > UnknownHostExceptions, even though /etc/hosts now contains host information > for it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10857) YarnClient Caching Addresses
[ https://issues.apache.org/jira/browse/YARN-10857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384426#comment-17384426 ] Steve Suh commented on YARN-10857: -- An unresolved InetSocketAddress object is created during the initialization of the YarnClient. This InetSocketAddress object is passed around and reused by the IPC [Client.java|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java] Here the InetSocketAddress is only checked if it's an unresolved address. It does not attempt to re-resolve the cached address and throws and error. https://github.com/apache/hadoop/blob/de41ce8a16434aee13f705a9e3666f29e8ec8cb3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Client.java#L1602-L1608 {code:java} if (address.isUnresolved()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} A possible fix would be to change this check and include the following: _if (address.isUnresolved() *{color:#4C9AFF}&& !updateAddress(){color}*)_ {code:java} if (address.isUnresolved() && !updateAddress()) { throw NetUtils.wrapException(address.getHostName(), address.getPort(), null, 0, new UnknownHostException()); } {code} > YarnClient Caching Addresses > > > Key: YARN-10857 > URL: https://issues.apache.org/jira/browse/YARN-10857 > Project: Hadoop YARN > Issue Type: Improvement > Components: client, yarn >Reporter: Steve Suh >Assignee: Prabhu Joseph >Priority: Minor > > We have noticed that when the YarnClient is initialized and used, it is not > very resilient when dns or /etc/hosts is modified in the following scenario: > Take for instance the following (and reproducable) sequence of events that > can occur on a service that instantiates and uses YarnClient. > - Yarn has rm HA enabled (*yarn.resourcemanager.ha.enabled* is *true*) and > there are two rms (rm1 and rm2). > - *yarn.client.failover-proxy-provider* is set to > *org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider* > 1)rm2 is currently the active rm > 2)/etc/hosts (or dns) is missing host information for rm2 > 3)A service is started and it initializes the YarnClient at startup. > 4)At some point in time after YarnClient is done initializing, /etc/hosts > is updated and contains host information for rm2 > 5)Yarn is queried, for instance calling *yarnclient.getApplications()* > 6)All YarnClient attempts to communicate with rm2 fail with > UnknownHostExceptions, even though /etc/hosts now contains host information > for it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10657) We should make max application per queue to support node label.
[ https://issues.apache.org/jira/browse/YARN-10657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384385#comment-17384385 ] Hadoop QA commented on YARN-10657: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 35s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:blue}0{color} | {color:blue} markdownlint {color} | {color:blue} 0m 0s{color} | {color:blue}{color} | {color:blue} markdownlint was not available. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} {color} | {color:green} 0m 0s{color} | {color:green}test4tests{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 28s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 35s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 9m 22s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 7s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 51s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 52s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 18s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 35s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 32s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 23m 6s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 0m 38s{color} | {color:blue}{color} | {color:blue} branch/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site no spotbugs output file (spotbugsXml.xml) {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 28s{color} | {color:blue}{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 5s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 40s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 40s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 0s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 8m 0s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 1m 49s{color} | {color:orange}https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/1131/artifact/out/diff-checkstyle-hadoop-yarn-project_hadoop-yarn.txt{color} | {color:orange} hadoop-yarn-project/hadoop-yarn: The patch generated 14 new + 523 unchanged - 0 fixed = 537 total (was 523) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green
[jira] [Commented] (YARN-10854) Support marking inactive node as untracked without configured include path
[ https://issues.apache.org/jira/browse/YARN-10854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384386#comment-17384386 ] Kuhu Shukla commented on YARN-10854: Proposal seems good but since I have been away from the land of YARN for a while, could [~brahma], [~templedf] or others chime in on the idea as well? I would love to review the code for this change. > Support marking inactive node as untracked without configured include path > -- > > Key: YARN-10854 > URL: https://issues.apache.org/jira/browse/YARN-10854 > Project: Hadoop YARN > Issue Type: Improvement > Components: resourcemanager >Reporter: Tao Yang >Assignee: Tao Yang >Priority: Major > Attachments: YARN-10854.001.patch > > > Currently inactive nodes which have been decommissioned/shutdown/lost for a > while(specified expiration time defined via > {{yarn.resourcemanager.node-removal-untracked.timeout-ms}}, 60 seconds by > default) and not exist in both include and exclude files can be marked as > untracked nodes and can be removed from RM state (YARN-4311). It's very > useful when auto-scaling is enabled in elastic cloud environment, which can > avoid unlimited increase of inactive nodes (mostly are decommissioned nodes). > But this only works when the include path is configured, mismatched for most > of our cloud environments without configured white list of nodes, which can > lead to easily control for the auto-scaling of nodes without further security > requirements. > So I propose to support marking inactive node as untracked without configured > include path, to be compatible with the former versions, we can add a switch > config for this. > Any thoughts/suggestions/feedbacks are welcome! -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-10856) Prevent ATS v2 health check REST API call if the ATS service itself is disabled.
[ https://issues.apache.org/jira/browse/YARN-10856?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benjamin Teke reassigned YARN-10856: Assignee: Benjamin Teke > Prevent ATS v2 health check REST API call if the ATS service itself is > disabled. > > > Key: YARN-10856 > URL: https://issues.apache.org/jira/browse/YARN-10856 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn-ui-v2 >Affects Versions: 3.3.1 >Reporter: Siddharth Ahuja >Assignee: Benjamin Teke >Priority: Major > > Currently, even if {{yarn.timeline-service.enabled}} is disabled, the UI2 > code still goes ahead and performs timeline health check REST API calls, see > [1], [2] and [3]. > This is un-necessary and can cause slowness issues with RM UI2 page loading > if a firewall is dropping packets on the ATS v2 port (as it is not meant to > be available e.g. 8188/8190) in the background and the timeout is not yet hit. > This ATSv2 health check REST API call is redundant and should be prevented if > the service itself is disabled. > [1] > https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/adapters/timeline-health.js > [2] > https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/routes/application.js#L34 > [3] > https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/routes/yarn-app/logs.js#L40 -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10657) We should make max application per queue to support node label.
[ https://issues.apache.org/jira/browse/YARN-10657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384245#comment-17384245 ] Andras Gyori commented on YARN-10657: - Thank you [~zhuqi] for the review. I agree with you that choosing the highest maximum application value is the least surprising behaviour for users. I believe that it is the safest and simplest approach to this problem. I have removed the default-node-label-expression based solution. > We should make max application per queue to support node label. > --- > > Key: YARN-10657 > URL: https://issues.apache.org/jira/browse/YARN-10657 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Qi Zhu >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10657.001.patch, YARN-10657.002.patch, > YARN-10657.003.patch, YARN-10657.004.patch > > > https://issues.apache.org/jira/browse/YARN-10641?focusedCommentId=17291708&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17291708 > As we discussed in above comment: > We should deep into the label related max applications per queue. > I think when node label enabled in queue, max applications should consider > the max capacity of all labels. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10657) We should make max application per queue to support node label.
[ https://issues.apache.org/jira/browse/YARN-10657?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andras Gyori updated YARN-10657: Attachment: YARN-10657.004.patch > We should make max application per queue to support node label. > --- > > Key: YARN-10657 > URL: https://issues.apache.org/jira/browse/YARN-10657 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Qi Zhu >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10657.001.patch, YARN-10657.002.patch, > YARN-10657.003.patch, YARN-10657.004.patch > > > https://issues.apache.org/jira/browse/YARN-10641?focusedCommentId=17291708&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17291708 > As we discussed in above comment: > We should deep into the label related max applications per queue. > I think when node label enabled in queue, max applications should consider > the max capacity of all labels. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-6221) Entities missing from ATS when summary log file info got returned to the ATS before the domain log
[ https://issues.apache.org/jira/browse/YARN-6221?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384227#comment-17384227 ] Hadoop QA commented on YARN-6221: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 53s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 1s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} {color} | {color:green} 0m 0s{color} | {color:green}test4tests{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 26m 44s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 29s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 31s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 18m 27s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 30s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 28s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 20m 9s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:green}+1{color} | {color:green} spotbugs {color} | {color:green} 0m 44s{color} | {color:green}{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 25s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 24s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:red}-1{color} | {color:red} javac {color} | {color:red} 0m 24s{color} | {color:red}https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/1130/artifact/out/diff-compile-javac-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timeline-pluginstorage-jdkUbuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04.txt{color} | {color:red} hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timeline-pluginstorage-jdkUbuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 20s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:red}-1{color} | {color:red} javac {color} | {color:red} 0m 20s{color} | {color:red}https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/1130/artifact/out/diff-compile-javac-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timeline-pluginstorage-jdkPrivateBuild-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10.txt{color} | {color:red} hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timeline-pluginstorage-jdkPrivateBuild-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 generated 1 new + 0 unchanged - 0 fixed = 1 total (was 0) {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 15s{color} | {color:ora
[jira] [Commented] (YARN-10867) YARN should expose a ENV used to map a custom device into docker container
[ https://issues.apache.org/jira/browse/YARN-10867?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384155#comment-17384155 ] Qi Zhu commented on YARN-10867: --- cc [~ebadger] [~epayne] [~gandras] [~Jim_Brennan] Could you take a look at this issue when you are free? Thanks. > YARN should expose a ENV used to map a custom device into docker container > -- > > Key: YARN-10867 > URL: https://issues.apache.org/jira/browse/YARN-10867 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Chi Heng >Priority: Major > > In some scenarios, like mounting a FUSE in docker,user needs to map a custom > device (eg. /dev/fuse) into docker container.I notice that an adddevice > method is defined in [ > hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java > > |https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] > ,I suppose that an ENV or config property should to be exposed to user to > call this method -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10727) ParentQueue does not validate the queue on removal
[ https://issues.apache.org/jira/browse/YARN-10727?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384095#comment-17384095 ] Hadoop QA commented on YARN-10727: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 7s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} {color} | {color:green} 0m 0s{color} | {color:green}test4tests{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 24m 24s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 12s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 57s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 49s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 59s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 47s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 46s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 44s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 20m 17s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:green}+1{color} | {color:green} spotbugs {color} | {color:green} 2m 1s{color} | {color:green}{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 55s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 0s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 0s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 52s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 52s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 44s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 52s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 15m 28s{color} | {color:green}{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 41s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 38s{color} | {color:green}{color} | {color:green} the
[jira] [Commented] (YARN-10838) Implement an optimised version of Configuration getPropsWithPrefix
[ https://issues.apache.org/jira/browse/YARN-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17384076#comment-17384076 ] Hadoop QA commented on YARN-10838: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Logfile || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 40s{color} | {color:blue}{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || || | {color:green}+1{color} | {color:green} dupname {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} No case conflicting files found. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} {color} | {color:green} 0m 0s{color} | {color:green}test4tests{color} | {color:green} The patch appears to include 3 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 27s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 1s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 51s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 58s{color} | {color:green}{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 58s{color} | {color:green}{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 46s{color} | {color:green}{color} | {color:green} trunk passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 43s{color} | {color:green}{color} | {color:green} trunk passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue} 18m 18s{color} | {color:blue}{color} | {color:blue} Both FindBugs and SpotBugs are enabled, using SpotBugs. {color} | | {color:green}+1{color} | {color:green} spotbugs {color} | {color:green} 1m 51s{color} | {color:green}{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 51s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 55s{color} | {color:green}{color} | {color:green} the patch passed with JDK Ubuntu-11.0.11+9-Ubuntu-0ubuntu2.20.04 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 55s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 45s{color} | {color:green}{color} | {color:green} the patch passed with JDK Private Build-1.8.0_292-8u292-b10-0ubuntu1~20.04-b10 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 45s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 42s{color} | {color:orange}https://ci-hadoop.apache.org/job/PreCommit-YARN-Build/1128/artifact/out/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 1 new + 109 unchanged - 0 fixed = 110 total (was 109) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 51s{color} | {color:green}{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green}{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 48s{color} | {color:green}{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {colo
[jira] [Assigned] (YARN-6221) Entities missing from ATS when summary log file info got returned to the ATS before the domain log
[ https://issues.apache.org/jira/browse/YARN-6221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rudolf Reti reassigned YARN-6221: - Assignee: Xiaomin Zhang (was: Li Lu) > Entities missing from ATS when summary log file info got returned to the ATS > before the domain log > -- > > Key: YARN-6221 > URL: https://issues.apache.org/jira/browse/YARN-6221 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn >Reporter: Sushmitha Sreenivasan >Assignee: Xiaomin Zhang >Priority: Critical > > Events data missing for the following entities: > curl -k --negotiate -u: > http://:8188/ws/v1/timeline/TEZ_APPLICATION_ATTEMPT/tez_appattempt_1487706062210_0012_01 > {"events":[],"entitytype":"TEZ_APPLICATION_ATTEMPT","entity":"tez_appattempt_1487706062210_0012_01","starttime":1487711606077,"domain":"Tez_ATS_application_1487706062210_0012","relatedentities":{"TEZ_DAG_ID":["dag_1487706062210_0012_2","dag_1487706062210_0012_1"]},"primaryfilters":{},"otherinfo":{}} > {code:title=Timeline Server log entry} > WARN timeline.TimelineDataManager > (TimelineDataManager.java:doPostEntities(366)) - Skip the timeline entity: { > id: tez_application_1487706062210_0012, type: TEZ_APPLICATION } > org.apache.hadoop.yarn.exceptions.YarnException: Domain information of the > timeline entity { id: tez_application_1487706062210_0012, type: > TEZ_APPLICATION } doesn't exist. > at > org.apache.hadoop.yarn.server.timeline.security.TimelineACLsManager.checkAccess(TimelineACLsManager.java:122) > at > org.apache.hadoop.yarn.server.timeline.TimelineDataManager.doPostEntities(TimelineDataManager.java:356) > at > org.apache.hadoop.yarn.server.timeline.TimelineDataManager.postEntities(TimelineDataManager.java:316) > at > org.apache.hadoop.yarn.server.timeline.EntityLogInfo.doParse(LogInfo.java:204) > at > org.apache.hadoop.yarn.server.timeline.LogInfo.parsePath(LogInfo.java:156) > at > org.apache.hadoop.yarn.server.timeline.LogInfo.parseForStore(LogInfo.java:113) > at > org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimelineStore$AppLogs.parseSummaryLogs(EntityGroupFSTimelineStore.java:682) > at > org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimelineStore$AppLogs.parseSummaryLogs(EntityGroupFSTimelineStore.java:657) > at > org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimelineStore$ActiveLogParser.run(EntityGroupFSTimelineStore.java:870) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10867) YARN should expose a ENV used to map a custom device into docker container
[ https://issues.apache.org/jira/browse/YARN-10867?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chi Heng updated YARN-10867: Description: In some scenarios, like mounting a FUSE in docker,user needs to map a custom device (eg. /dev/fuse) into docker container.I notice that an adddevice method is defined in [ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java |https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] ,I suppose that an ENV or config property should to be exposed to user to call this method (was: In some scenarios, like mounting a FUSE in docker,user needs to map a custom device (eg. /dev/fuse) into docker container.I notice that an adddevice method is defined in [ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java |http://gitlab.qiyi.domain/bigdata/hadoop/blob/branch-3.2.2/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] ,I suppose that an ENV or config property should to be exposed to user to call this method) > YARN should expose a ENV used to map a custom device into docker container > -- > > Key: YARN-10867 > URL: https://issues.apache.org/jira/browse/YARN-10867 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Chi Heng >Priority: Major > > In some scenarios, like mounting a FUSE in docker,user needs to map a custom > device (eg. /dev/fuse) into docker container.I notice that an adddevice > method is defined in [ > hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java > > |https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] > ,I suppose that an ENV or config property should to be exposed to user to > call this method -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-10867) YARN should expose a ENV used to map a custom device into docker container
Chi Heng created YARN-10867: --- Summary: YARN should expose a ENV used to map a custom device into docker container Key: YARN-10867 URL: https://issues.apache.org/jira/browse/YARN-10867 Project: Hadoop YARN Issue Type: Improvement Reporter: Chi Heng In some scenarios, like mounting a FUSE in docker,user needs to map a custom device (eg. /dev/fuse) into docker container.I notice that an adddevice method is defined in [ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java |http://gitlab.qiyi.domain/bigdata/hadoop/blob/branch-3.2.2/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/runtime/docker/DockerRunCommand.java] ,I suppose that an ENV or config property should to be exposed to user to call this method -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-10857) YarnClient Caching Addresses
[ https://issues.apache.org/jira/browse/YARN-10857?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Prabhu Joseph reassigned YARN-10857: Assignee: Prabhu Joseph > YarnClient Caching Addresses > > > Key: YARN-10857 > URL: https://issues.apache.org/jira/browse/YARN-10857 > Project: Hadoop YARN > Issue Type: Improvement > Components: client, yarn >Reporter: Steve Suh >Assignee: Prabhu Joseph >Priority: Minor > > We have noticed that when the YarnClient is initialized and used, it is not > very resilient when dns or /etc/hosts is modified in the following scenario: > Take for instance the following (and reproducable) sequence of events that > can occur on a service that instantiates and uses YarnClient. > - Yarn has rm HA enabled (*yarn.resourcemanager.ha.enabled* is *true*) and > there are two rms (rm1 and rm2). > - *yarn.client.failover-proxy-provider* is set to > *org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider* > 1)rm2 is currently the active rm > 2)/etc/hosts (or dns) is missing host information for rm2 > 3)A service is started and it initializes the YarnClient at startup. > 4)At some point in time after YarnClient is done initializing, /etc/hosts > is updated and contains host information for rm2 > 5)Yarn is queried, for instance calling *yarnclient.getApplications()* > 6)All YarnClient attempts to communicate with rm2 fail with > UnknownHostExceptions, even though /etc/hosts now contains host information > for it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10727) ParentQueue does not validate the queue on removal
[ https://issues.apache.org/jira/browse/YARN-10727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andras Gyori updated YARN-10727: Attachment: (was: YARN-10727.001.patch) > ParentQueue does not validate the queue on removal > -- > > Key: YARN-10727 > URL: https://issues.apache.org/jira/browse/YARN-10727 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10727.001.patch > > > With the addition of YARN-10532 ParentQueue has a public method, removeQueue, > which allows the deletion of a queue at runtime. However, there is no > validation regarding the queue which is to be removed, therefore it is > possible to remove a queue from the CSQueueManager that is not a child of the > ParentQueue. Since it is a public method, there must be validations such as: > * check, if the parent of the queue to be removed is the current ParentQueue > * check, if the parent actually contains the queue in its childQueues > collection -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10727) ParentQueue does not validate the queue on removal
[ https://issues.apache.org/jira/browse/YARN-10727?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andras Gyori updated YARN-10727: Attachment: YARN-10727.001.patch > ParentQueue does not validate the queue on removal > -- > > Key: YARN-10727 > URL: https://issues.apache.org/jira/browse/YARN-10727 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10727.001.patch > > > With the addition of YARN-10532 ParentQueue has a public method, removeQueue, > which allows the deletion of a queue at runtime. However, there is no > validation regarding the queue which is to be removed, therefore it is > possible to remove a queue from the CSQueueManager that is not a child of the > ParentQueue. Since it is a public method, there must be validations such as: > * check, if the parent of the queue to be removed is the current ParentQueue > * check, if the parent actually contains the queue in its childQueues > collection -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10838) Implement an optimised version of Configuration getPropsWithPrefix
[ https://issues.apache.org/jira/browse/YARN-10838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andras Gyori updated YARN-10838: Attachment: YARN-10838.002.patch > Implement an optimised version of Configuration getPropsWithPrefix > -- > > Key: YARN-10838 > URL: https://issues.apache.org/jira/browse/YARN-10838 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > Attachments: YARN-10838.001.patch, YARN-10838.002.patch > > > AutoCreatedQueueTemplate also has multiple call to > Configuration#getPropsWithPrefix. It must be eliminated in order to improve > the performance on reinitialisation. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10866) RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing
[ https://issues.apache.org/jira/browse/YARN-10866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Prabhu Joseph updated YARN-10866: - Description: RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing in /etc/hosts {code} 2021-07-19 13:07:18,892 ERROR [Listener at 0.0.0.0/45951] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting MRAppMaster java.lang.IllegalArgumentException: java.net.UnknownHostException: resourcemanager-1.resourcemanager at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:466) at org.apache.hadoop.yarn.client.ClientRMProxy.getTokenService(ClientRMProxy.java:154) at org.apache.hadoop.yarn.client.ClientRMProxy.getAMRMTokenService(ClientRMProxy.java:139) at org.apache.hadoop.yarn.client.ClientRMProxy.setAMRMTokenService(ClientRMProxy.java:81) at org.apache.hadoop.yarn.client.ClientRMProxy.getRMAddress(ClientRMProxy.java:100) at org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider.getProxyInternal(ConfiguredRMFailoverProxyProvider.java:76) at org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider.init(RequestHedgingRMFailoverProxyProvider.java:75) at org.apache.hadoop.yarn.client.RMProxy.createRMFailoverProxyProvider(RMProxy.java:194) at org.apache.hadoop.yarn.client.RMProxy.newProxyInstance(RMProxy.java:130) at org.apache.hadoop.yarn.client.RMProxy.createRMProxy(RMProxy.java:103) {code} was: RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing {code} 2021-07-19 13:07:18,892 ERROR [Listener at 0.0.0.0/45951] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting MRAppMaster java.lang.IllegalArgumentException: java.net.UnknownHostException: resourcemanager-1.resourcemanager at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:466) at org.apache.hadoop.yarn.client.ClientRMProxy.getTokenService(ClientRMProxy.java:154) at org.apache.hadoop.yarn.client.ClientRMProxy.getAMRMTokenService(ClientRMProxy.java:139) at org.apache.hadoop.yarn.client.ClientRMProxy.setAMRMTokenService(ClientRMProxy.java:81) at org.apache.hadoop.yarn.client.ClientRMProxy.getRMAddress(ClientRMProxy.java:100) at org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider.getProxyInternal(ConfiguredRMFailoverProxyProvider.java:76) at org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider.init(RequestHedgingRMFailoverProxyProvider.java:75) at org.apache.hadoop.yarn.client.RMProxy.createRMFailoverProxyProvider(RMProxy.java:194) at org.apache.hadoop.yarn.client.RMProxy.newProxyInstance(RMProxy.java:130) at org.apache.hadoop.yarn.client.RMProxy.createRMProxy(RMProxy.java:103) {code} > RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if > standby host info is missing > --- > > Key: YARN-10866 > URL: https://issues.apache.org/jira/browse/YARN-10866 > Project: Hadoop YARN > Issue Type: Bug > Components: client >Affects Versions: 3.3.1 >Reporter: Prabhu Joseph >Assignee: Prabhu Joseph >Priority: Major > > RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if > standby host info is missing in /etc/hosts > {code} > 2021-07-19 13:07:18,892 ERROR [Listener at 0.0.0.0/45951] > org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting MRAppMaster > java.lang.IllegalArgumentException: java.net.UnknownHostException: > resourcemanager-1.resourcemanager > at > org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:466) > at > org.apache.hadoop.yarn.client.ClientRMProxy.getTokenService(ClientRMProxy.java:154) > at > org.apache.hadoop.yarn.client.ClientRMProxy.getAMRMTokenService(ClientRMProxy.java:139) > at > org.apache.hadoop.yarn.client.ClientRMProxy.setAMRMTokenService(ClientRMProxy.java:81) > at > org.apache.hadoop.yarn.client.ClientRMProxy.getRMAddress(ClientRMProxy.java:100) > at > org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider.getProxyInternal(ConfiguredRMFailoverProxyProvider.java:76) > at > org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider.init(RequestHedgingRMFailoverProxyProvider.java:75) > at > org.apache.hadoop.yarn.client.RMProxy.createRMFailoverProxyProvider(RMProxy.java:194) > at > org.apache.hadoop.yarn.client.RMProxy.newProxyInstance(RMProxy.java:130) > at > org.apache.hadoop.yarn.client.RMProxy.createRMProxy(RMProxy.java:103) > {code} -- This message was sent by Atlassian Jira (v8.3.
[jira] [Created] (YARN-10866) RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing
Prabhu Joseph created YARN-10866: Summary: RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing Key: YARN-10866 URL: https://issues.apache.org/jira/browse/YARN-10866 Project: Hadoop YARN Issue Type: Bug Components: client Affects Versions: 3.3.1 Reporter: Prabhu Joseph Assignee: Prabhu Joseph RequestHedgingRMFailoverProxyProvider fails to connect to Active RM if standby host info is missing {code} 2021-07-19 13:07:18,892 ERROR [Listener at 0.0.0.0/45951] org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Error starting MRAppMaster java.lang.IllegalArgumentException: java.net.UnknownHostException: resourcemanager-1.resourcemanager at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:466) at org.apache.hadoop.yarn.client.ClientRMProxy.getTokenService(ClientRMProxy.java:154) at org.apache.hadoop.yarn.client.ClientRMProxy.getAMRMTokenService(ClientRMProxy.java:139) at org.apache.hadoop.yarn.client.ClientRMProxy.setAMRMTokenService(ClientRMProxy.java:81) at org.apache.hadoop.yarn.client.ClientRMProxy.getRMAddress(ClientRMProxy.java:100) at org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider.getProxyInternal(ConfiguredRMFailoverProxyProvider.java:76) at org.apache.hadoop.yarn.client.RequestHedgingRMFailoverProxyProvider.init(RequestHedgingRMFailoverProxyProvider.java:75) at org.apache.hadoop.yarn.client.RMProxy.createRMFailoverProxyProvider(RMProxy.java:194) at org.apache.hadoop.yarn.client.RMProxy.newProxyInstance(RMProxy.java:130) at org.apache.hadoop.yarn.client.RMProxy.createRMProxy(RMProxy.java:103) {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10865) Add UI support to display NodeHealthDetails in NodePage
[ https://issues.apache.org/jira/browse/YARN-10865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10865: - Attachment: (was: image-2021-07-20-12-42-59-013.png) > Add UI support to display NodeHealthDetails in NodePage > --- > > Key: YARN-10865 > URL: https://issues.apache.org/jira/browse/YARN-10865 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Major > Attachments: image-2021-07-20-12-43-41-273.png > > > Add support to display NodeHealthDetails resource information on Node Page. > Sample Page: > !image-2021-07-20-12-43-41-273.png|width=1423,height=250! -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10865) Add UI support to display NodeHealthDetails in NodePage
[ https://issues.apache.org/jira/browse/YARN-10865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10865: - Attachment: image-2021-07-20-12-43-41-273.png > Add UI support to display NodeHealthDetails in NodePage > --- > > Key: YARN-10865 > URL: https://issues.apache.org/jira/browse/YARN-10865 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Major > Attachments: image-2021-07-20-12-42-59-013.png, > image-2021-07-20-12-43-41-273.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10865) Add UI support to display NodeHealthDetails in NodePage
[ https://issues.apache.org/jira/browse/YARN-10865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10865: - Description: Add support to display NodeHealthDetails resource information on Node Page. Sample Page: !image-2021-07-20-12-43-41-273.png|width=1423,height=250! > Add UI support to display NodeHealthDetails in NodePage > --- > > Key: YARN-10865 > URL: https://issues.apache.org/jira/browse/YARN-10865 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Major > Attachments: image-2021-07-20-12-42-59-013.png, > image-2021-07-20-12-43-41-273.png > > > Add support to display NodeHealthDetails resource information on Node Page. > Sample Page: > !image-2021-07-20-12-43-41-273.png|width=1423,height=250! -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10865) Add UI support to display NodeHealthDetails in NodePage
[ https://issues.apache.org/jira/browse/YARN-10865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Cyrus Jackson updated YARN-10865: - Attachment: image-2021-07-20-12-42-59-013.png > Add UI support to display NodeHealthDetails in NodePage > --- > > Key: YARN-10865 > URL: https://issues.apache.org/jira/browse/YARN-10865 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Cyrus Jackson >Assignee: Cyrus Jackson >Priority: Major > Attachments: image-2021-07-20-12-42-59-013.png, > image-2021-07-20-12-43-41-273.png > > -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Resolved] (YARN-10859) [UI2] Extend Applications with queue path instead of queue name
[ https://issues.apache.org/jira/browse/YARN-10859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andras Gyori resolved YARN-10859. - Resolution: Not A Problem > [UI2] Extend Applications with queue path instead of queue name > --- > > Key: YARN-10859 > URL: https://issues.apache.org/jira/browse/YARN-10859 > Project: Hadoop YARN > Issue Type: Improvement > Components: capacity scheduler, yarn-ui-v2 >Reporter: Andras Gyori >Assignee: Andras Gyori >Priority: Major > > YARN-10630 discovered a bug in UI2, where ambiguous queues (root.a.b <-> > root.b) could not be handled properly. It would also eliminate any further > inconsistency if we replaced queue names with queue paths. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org