[jira] [Updated] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5773: --- Attachment: YARN-5773.0005.patch Ignore Yarn-5773.0004 patch. Attaching latest patch > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.0004.patch, YARN-5773.0005.patch, YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4363) In TestFairScheduler, testcase should not create FairScheduler redundantly
[ https://issues.apache.org/jira/browse/YARN-4363?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610897#comment-15610897 ] Hudson commented on YARN-4363: -- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10695 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10695/]) YARN-4363. In TestFairScheduler, testcase should not create (rohithsharmaks: rev e29cba61a0bec0629de287f67ae6eed526295ffb) * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairScheduler.java > In TestFairScheduler, testcase should not create FairScheduler redundantly > -- > > Key: YARN-4363 > URL: https://issues.apache.org/jira/browse/YARN-4363 > Project: Hadoop YARN > Issue Type: Test > Components: fairscheduler >Affects Versions: 2.6.0 >Reporter: Tao Jie >Assignee: Tao Jie >Priority: Trivial > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: YARN-4363.0002.patch, YARN-4363.001.patch > > > I am trying to make some improvement on fairscheduler, but get some test > failure on TestFairScheduler, due to redundant FairScheduler creation: > In TestFairScheduler, FairScheduler and RM is created, then set RMContext of > RM to scheduler. > {code} > @Before > public void setUp() throws IOException { > scheduler = new FairScheduler(); > conf = createConfiguration(); > resourceManager = new MockRM(conf); > scheduler.setRMContext(resourceManager.getRMContext()); > } > {code} > However in several case, scheduler is renewed, as a result RMcontext in > scheduler is null. > {code} > @Test > public void testMinZeroResourcesSettings() throws IOException { > scheduler = new FairScheduler(); > YarnConfiguration conf = new YarnConfiguration(); > ... > scheduler.init(conf); > {code} > Then do scheduler.init(conf), I get a NPE(I try to get something from > RMContext in scheduler initialization). > So FairScheduler should not be renewed in test block. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610886#comment-15610886 ] Bibin A Chundatt commented on YARN-5773: [~varun_saxena] Earlier implementation also followed the same. if i understand correctly > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.0004.patch, YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610879#comment-15610879 ] Varun Saxena commented on YARN-5773: [~bibinchundatt], Below check for getNumActiveApplications is not required. If cluster resources are 0, there is no point activating even one app. {code} 703 if (!Resources.greaterThan(resourceCalculator, lastClusterResource, 704 lastClusterResource, Resources.none()) 705 && !(getNumActiveApplications() < 1)) { 706 return; 707 } {code} I think a log can be added under the condition, maybe at DEBUG log level to avoid too many logs. bq. Cluster resource UI is self explanatory, so required be add ?? I am +0 on this. [~sunilg], your thoughts on this ? > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.0004.patch, YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610875#comment-15610875 ] Bibin A Chundatt commented on YARN-5773: [~varun_saxena] Thank you for review comment. Attached latest patch handling all comments. > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.0004.patch, YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4555) TestDefaultContainerExecutor#testContainerLaunchError fails on non-english locale environment
[ https://issues.apache.org/jira/browse/YARN-4555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610874#comment-15610874 ] Hadoop QA commented on YARN-4555: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 18s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 19s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 48s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 16s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 18s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 15m 41s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 16s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 31m 42s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-4555 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12780967/YARN-4555.1.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 2346e7e50ab8 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 9f32364 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13533/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13533/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > TestDefaultContainerExecutor#testContainerLaunchError fails on non-english > locale environment > - > > Key: YARN-4555 > URL: https://issues.apache.org/jira/browse/YARN-4555 > Project: Hadoop YARN > Issue Type: Test > Components: nodemanager, test >Affects Versions: 2.7.1 >Reporter: Takashi Ohnishi >Assignee: Takashi Ohn
[jira] [Updated] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5773: --- Attachment: YARN-5773.0004.patch > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.0004.patch, YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5428) Allow for specifying the docker client configuration directory
[ https://issues.apache.org/jira/browse/YARN-5428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610864#comment-15610864 ] Vinod Kumar Vavilapalli commented on YARN-5428: --- [~tangzhankun], there is an enthusiastic community working on the Docker / YARN pieces, we welcome your contributions and feedback. Tx. > Allow for specifying the docker client configuration directory > -- > > Key: YARN-5428 > URL: https://issues.apache.org/jira/browse/YARN-5428 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn >Reporter: Shane Kumpf >Assignee: Shane Kumpf > Attachments: YARN-5428.001.patch, YARN-5428.002.patch, > YARN-5428.003.patch, YARN-5428.004.patch, > YARN-5428Allowforspecifyingthedockerclientconfigurationdirectory.pdf > > > The docker client allows for specifying a configuration directory that > contains the docker client's configuration. It is common to store "docker > login" credentials in this config, to avoid the need to docker login on each > cluster member. > By default the docker client config is $HOME/.docker/config.json on Linux. > However, this does not work with the current container executor user > switching and it may also be desirable to centralize this configuration > beyond the single user's home directory. > Note that the command line arg is for the configuration directory NOT the > configuration file. > This change will be needed to allow YARN to automatically pull images at > localization time or within container executor. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4363) In TestFairScheduler, testcase should not create FairScheduler redundantly
[ https://issues.apache.org/jira/browse/YARN-4363?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610853#comment-15610853 ] Hadoop QA commented on YARN-4363: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 39s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 21s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 22s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 15s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 8s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 36m 7s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 52m 13s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-4363 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835480/YARN-4363.0002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux ee2a8534bd3b 3.13.0-93-generic #140-Ubuntu SMP Mon Jul 18 21:21:05 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 9f32364 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13531/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13531/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > In TestFairScheduler, testcase should not create FairScheduler redundantly > -- > > Key: YARN-4363 > URL: https://issues.apache.org/jira/browse/YARN-4363 > Project: Hadoop YARN > Issue Type: Test > Components: fairscheduler >Affects Versions: 2.6.0 >Reporter: Tao Jie >Assignee: Tao Jie >Priority: Trivial > Attachmen
[jira] [Commented] (YARN-4555) TestDefaultContainerExecutor#testContainerLaunchError fails on non-english locale environment
[ https://issues.apache.org/jira/browse/YARN-4555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610851#comment-15610851 ] Hudson commented on YARN-4555: -- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10694 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10694/]) YARN-4555. TestDefaultContainerExecutor#testContainerLaunchError fails (rohithsharmaks: rev b110c4b5e82d8310f13a22bba1c8afbcca80144f) * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/TestDefaultContainerExecutor.java > TestDefaultContainerExecutor#testContainerLaunchError fails on non-english > locale environment > - > > Key: YARN-4555 > URL: https://issues.apache.org/jira/browse/YARN-4555 > Project: Hadoop YARN > Issue Type: Test > Components: nodemanager, test >Affects Versions: 2.7.1 >Reporter: Takashi Ohnishi >Assignee: Takashi Ohnishi >Priority: Minor > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: YARN-4555.1.patch > > > In my env where LANG=ja_JP.UTF-8, the test fails with > {code} > --- > Test set: > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > --- > Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 1.286 sec <<< > FAILURE! - in > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > testContainerLaunchError(org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor) > Time elapsed: 1.149 sec <<< FAILURE! > java.lang.AssertionError: Invalid Diagnostics message: Exception from > container-launch. > Container id: CONTAINER_ID > Exit code: 127 > Exception message: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディレクトリはありません > Stack trace: ExitCodeException exitCode=127: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディ>レクトリはありません > {code} > This is because the test code assertion assumes the English locale as below. > {code} > 250 public Object answer(InvocationOnMock invocationOnMock) > 251 throws Throwable { > 252 String diagnostics = (String) > invocationOnMock.getArguments()[0]; > 253 assertTrue("Invalid Diagnostics message: " + diagnostics, > 254 diagnostics.contains("No such file or directory")); > 255 return null; > 256 } > {code} > This exists on trunk, too. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-5490) [YARN-3368] Fix various alignment issues in Node page and the broken breadcrumb link
[ https://issues.apache.org/jira/browse/YARN-5490?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB reassigned YARN-5490: -- Assignee: Akhil PB (was: Sunil G) > [YARN-3368] Fix various alignment issues in Node page and the broken > breadcrumb link > > > Key: YARN-5490 > URL: https://issues.apache.org/jira/browse/YARN-5490 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sunil G >Assignee: Akhil PB > > Few alignment issue in nodes page > breadcrumb view is not showing the node table when clicked on Nodes option. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-4555) TestDefaultContainerExecutor#testContainerLaunchError fails on non-english locale environment
[ https://issues.apache.org/jira/browse/YARN-4555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rohith Sharma K S updated YARN-4555: Summary: TestDefaultContainerExecutor#testContainerLaunchError fails on non-english locale environment (was: TestDefaultContainerExecutor#testContainerLaunchError fails on non-englihsh locale environment) > TestDefaultContainerExecutor#testContainerLaunchError fails on non-english > locale environment > - > > Key: YARN-4555 > URL: https://issues.apache.org/jira/browse/YARN-4555 > Project: Hadoop YARN > Issue Type: Test > Components: nodemanager, test >Affects Versions: 2.7.1 >Reporter: Takashi Ohnishi >Assignee: Takashi Ohnishi >Priority: Minor > Fix For: 2.9.0, 3.0.0-alpha2 > > Attachments: YARN-4555.1.patch > > > In my env where LANG=ja_JP.UTF-8, the test fails with > {code} > --- > Test set: > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > --- > Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 1.286 sec <<< > FAILURE! - in > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > testContainerLaunchError(org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor) > Time elapsed: 1.149 sec <<< FAILURE! > java.lang.AssertionError: Invalid Diagnostics message: Exception from > container-launch. > Container id: CONTAINER_ID > Exit code: 127 > Exception message: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディレクトリはありません > Stack trace: ExitCodeException exitCode=127: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディ>レクトリはありません > {code} > This is because the test code assertion assumes the English locale as below. > {code} > 250 public Object answer(InvocationOnMock invocationOnMock) > 251 throws Throwable { > 252 String diagnostics = (String) > invocationOnMock.getArguments()[0]; > 253 assertTrue("Invalid Diagnostics message: " + diagnostics, > 254 diagnostics.contains("No such file or directory")); > 255 return null; > 256 } > {code} > This exists on trunk, too. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5500) 'Master node' link under application tab is broken
[ https://issues.apache.org/jira/browse/YARN-5500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610821#comment-15610821 ] Hadoop QA commented on YARN-5500: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 6s{color} | {color:red} YARN-5500 does not apply to trunk. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | YARN-5500 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835486/YARN-5500.001.patch | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13534/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > 'Master node' link under application tab is broken > -- > > Key: YARN-5500 > URL: https://issues.apache.org/jira/browse/YARN-5500 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sumana Sathish >Assignee: Akhil PB >Priority: Critical > Attachments: YARN-5500.001.patch > > > Steps to reproduce: > * Click on the running application portion on the donut under "Cluster > resource usage by applications" > * Under App Master Info, there is a link provided for "Master Node". > The link is broken. It doesn't redirect to any page. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5500) 'Master node' link under application tab is broken
[ https://issues.apache.org/jira/browse/YARN-5500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610806#comment-15610806 ] Akhil PB commented on YARN-5500: Uploaded YARN-5500.001.patch > 'Master node' link under application tab is broken > -- > > Key: YARN-5500 > URL: https://issues.apache.org/jira/browse/YARN-5500 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sumana Sathish >Assignee: Akhil PB >Priority: Critical > Attachments: YARN-5500.001.patch > > > Steps to reproduce: > * Click on the running application portion on the donut under "Cluster > resource usage by applications" > * Under App Master Info, there is a link provided for "Master Node". > The link is broken. It doesn't redirect to any page. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5500) 'Master node' link under application tab is broken
[ https://issues.apache.org/jira/browse/YARN-5500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB updated YARN-5500: --- Attachment: YARN-5500.001.patch > 'Master node' link under application tab is broken > -- > > Key: YARN-5500 > URL: https://issues.apache.org/jira/browse/YARN-5500 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sumana Sathish >Assignee: Akhil PB >Priority: Critical > Attachments: YARN-5500.001.patch > > > Steps to reproduce: > * Click on the running application portion on the donut under "Cluster > resource usage by applications" > * Under App Master Info, there is a link provided for "Master Node". > The link is broken. It doesn't redirect to any page. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4555) TestDefaultContainerExecutor#testContainerLaunchError fails on non-englihsh locale environment
[ https://issues.apache.org/jira/browse/YARN-4555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610798#comment-15610798 ] Rohith Sharma K S commented on YARN-4555: - committing shortly.. > TestDefaultContainerExecutor#testContainerLaunchError fails on non-englihsh > locale environment > -- > > Key: YARN-4555 > URL: https://issues.apache.org/jira/browse/YARN-4555 > Project: Hadoop YARN > Issue Type: Test > Components: nodemanager, test >Affects Versions: 2.7.1 >Reporter: Takashi Ohnishi >Assignee: Takashi Ohnishi >Priority: Minor > Attachments: YARN-4555.1.patch > > > In my env where LANG=ja_JP.UTF-8, the test fails with > {code} > --- > Test set: > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > --- > Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 1.286 sec <<< > FAILURE! - in > org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor > testContainerLaunchError(org.apache.hadoop.yarn.server.nodemanager.TestDefaultContainerExecutor) > Time elapsed: 1.149 sec <<< FAILURE! > java.lang.AssertionError: Invalid Diagnostics message: Exception from > container-launch. > Container id: CONTAINER_ID > Exit code: 127 > Exception message: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディレクトリはありません > Stack trace: ExitCodeException exitCode=127: bash: > target/TestDefaultContainerExecutor/localDir/default_container_executor.sh: > そのようなファイルやディ>レクトリはありません > {code} > This is because the test code assertion assumes the English locale as below. > {code} > 250 public Object answer(InvocationOnMock invocationOnMock) > 251 throws Throwable { > 252 String diagnostics = (String) > invocationOnMock.getArguments()[0]; > 253 assertTrue("Invalid Diagnostics message: " + diagnostics, > 254 diagnostics.contains("No such file or directory")); > 255 return null; > 256 } > {code} > This exists on trunk, too. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4734) Merge branch:YARN-3368 to trunk
[ https://issues.apache.org/jira/browse/YARN-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610773#comment-15610773 ] Hadoop QA commented on YARN-4734: - (!) A patch to the testing environment has been detected. Re-executing against the patched versions to perform further tests. The console is at https://builds.apache.org/job/PreCommit-YARN-Build/13532/console in case of problems. > Merge branch:YARN-3368 to trunk > --- > > Key: YARN-4734 > URL: https://issues.apache.org/jira/browse/YARN-4734 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-4734.1.patch, YARN-4734.10-NOT_READY.patch, > YARN-4734.11-NOT_READY.patch, YARN-4734.12-NOT_READY.patch, > YARN-4734.13.patch, YARN-4734.14.patch, YARN-4734.15.patch, > YARN-4734.16.patch, YARN-4734.17-rebased.patch, YARN-4734.17.patch, > YARN-4734.2.patch, YARN-4734.3.patch, YARN-4734.4.patch, YARN-4734.5.patch, > YARN-4734.6.patch, YARN-4734.7.patch, YARN-4734.8.patch, > YARN-4734.9-NOT_READY.patch > > > YARN-2928 branch is planned to merge back to trunk shortly, it depends on > changes of YARN-3368. This JIRA is to track the merging task. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-5500) 'Master node' link under application tab is broken
[ https://issues.apache.org/jira/browse/YARN-5500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB reassigned YARN-5500: -- Assignee: Akhil PB > 'Master node' link under application tab is broken > -- > > Key: YARN-5500 > URL: https://issues.apache.org/jira/browse/YARN-5500 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sumana Sathish >Assignee: Akhil PB >Priority: Critical > > Steps to reproduce: > * Click on the running application portion on the donut under "Cluster > resource usage by applications" > * Under App Master Info, there is a link provided for "Master Node". > The link is broken. It doesn't redirect to any page. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2599) Standby RM should also expose some jmx and metrics
[ https://issues.apache.org/jira/browse/YARN-2599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610762#comment-15610762 ] Naganarasimha G R commented on YARN-2599: - Hi [~kasha] & [~rohithsharma], Seems like we have not completely established what JMX and metrics which we are planning to expose from the standby RM so it would be ideal to come up with it before starting to expose the JMX and metrics from standby. Hence cancelling the patch. > Standby RM should also expose some jmx and metrics > -- > > Key: YARN-2599 > URL: https://issues.apache.org/jira/browse/YARN-2599 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager >Affects Versions: 2.5.1 >Reporter: Karthik Kambatla >Assignee: Rohith Sharma K S > Attachments: YARN-2599.patch > > > YARN-1898 redirects jmx and metrics to the Active. As discussed there, we > need to separate out metrics displayed so the Standby RM can also be > monitored. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-4734) Merge branch:YARN-3368 to trunk
[ https://issues.apache.org/jira/browse/YARN-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Wangda Tan updated YARN-4734: - Attachment: YARN-4734.17-rebased.patch Thanks [~aw], one very recent change in trunk causes conflict, just rebased. One question: does this mean in the future all the precommit will run on latest Yetus master? Or we need to trigger Jenkins with some parameters to run with latest Yetus? > Merge branch:YARN-3368 to trunk > --- > > Key: YARN-4734 > URL: https://issues.apache.org/jira/browse/YARN-4734 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-4734.1.patch, YARN-4734.10-NOT_READY.patch, > YARN-4734.11-NOT_READY.patch, YARN-4734.12-NOT_READY.patch, > YARN-4734.13.patch, YARN-4734.14.patch, YARN-4734.15.patch, > YARN-4734.16.patch, YARN-4734.17-rebased.patch, YARN-4734.17.patch, > YARN-4734.2.patch, YARN-4734.3.patch, YARN-4734.4.patch, YARN-4734.5.patch, > YARN-4734.6.patch, YARN-4734.7.patch, YARN-4734.8.patch, > YARN-4734.9-NOT_READY.patch > > > YARN-2928 branch is planned to merge back to trunk shortly, it depends on > changes of YARN-3368. This JIRA is to track the merging task. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2995) Enhance UI to show cluster resource utilization of various container types
[ https://issues.apache.org/jira/browse/YARN-2995?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610731#comment-15610731 ] Hadoop QA commented on YARN-2995: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 6 new or modified test files. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 32s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 40s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 54s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 31s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 15s{color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 6m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 51s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 1m 34s{color} | {color:orange} root: The patch generated 13 new + 289 unchanged - 3 fixed = 302 total (was 292) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 43s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 37s{color} | {color:green} hadoop-yarn-server-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 15m 36s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 36m 3s{color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 4s{color} | {color:green} hadoop-sls in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 27s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}120m 14s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.resourcemanager.TestRMRestart | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-2995 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835465/YARN-2995.002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle cc | | uname | Linux af516a94a785 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | tr
[jira] [Updated] (YARN-4363) In TestFairScheduler, testcase should not create FairScheduler redundantly
[ https://issues.apache.org/jira/browse/YARN-4363?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rohith Sharma K S updated YARN-4363: Attachment: YARN-4363.0002.patch +1 for the patch.. patch is stale, I just rebased against trunk and reattaching the patch.. I will commit patch once HadoopQA triggers. > In TestFairScheduler, testcase should not create FairScheduler redundantly > -- > > Key: YARN-4363 > URL: https://issues.apache.org/jira/browse/YARN-4363 > Project: Hadoop YARN > Issue Type: Test > Components: fairscheduler >Affects Versions: 2.6.0 >Reporter: Tao Jie >Assignee: Tao Jie >Priority: Trivial > Attachments: YARN-4363.0002.patch, YARN-4363.001.patch > > > I am trying to make some improvement on fairscheduler, but get some test > failure on TestFairScheduler, due to redundant FairScheduler creation: > In TestFairScheduler, FairScheduler and RM is created, then set RMContext of > RM to scheduler. > {code} > @Before > public void setUp() throws IOException { > scheduler = new FairScheduler(); > conf = createConfiguration(); > resourceManager = new MockRM(conf); > scheduler.setRMContext(resourceManager.getRMContext()); > } > {code} > However in several case, scheduler is renewed, as a result RMcontext in > scheduler is null. > {code} > @Test > public void testMinZeroResourcesSettings() throws IOException { > scheduler = new FairScheduler(); > YarnConfiguration conf = new YarnConfiguration(); > ... > scheduler.init(conf); > {code} > Then do scheduler.init(conf), I get a NPE(I try to get something from > RMContext in scheduler initialization). > So FairScheduler should not be renewed in test block. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5772) Replace old Hadoop logo with new one
[ https://issues.apache.org/jira/browse/YARN-5772?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sunil G updated YARN-5772: -- Release Note: (was: Thanks Akhil for the patch. And thank you Akira Ajisaka for raising and reviewing the patch!) > Replace old Hadoop logo with new one > > > Key: YARN-5772 > URL: https://issues.apache.org/jira/browse/YARN-5772 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn-ui-v2 >Affects Versions: YARN-3368 >Reporter: Akira Ajisaka >Assignee: Akhil PB > Attachments: YARN-5772-YARN-3368.0001.patch, ui2-with-newlogo.png > > > YARN-5161 added Apache Hadoop logo in the UI but the logo is old. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5772) Replace old Hadoop logo with new one
[ https://issues.apache.org/jira/browse/YARN-5772?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610711#comment-15610711 ] Sunil G commented on YARN-5772: --- Thanks [~akhilpb] for the patch. And thank you [~ajisakaa] for raising and reviewing the patch! > Replace old Hadoop logo with new one > > > Key: YARN-5772 > URL: https://issues.apache.org/jira/browse/YARN-5772 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn-ui-v2 >Affects Versions: YARN-3368 >Reporter: Akira Ajisaka >Assignee: Akhil PB > Attachments: YARN-5772-YARN-3368.0001.patch, ui2-with-newlogo.png > > > YARN-5161 added Apache Hadoop logo in the UI but the logo is old. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5705) [YARN-3368] Add support for Timeline V2 to new web UI
[ https://issues.apache.org/jira/browse/YARN-5705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB updated YARN-5705: --- Attachment: (was: YARN-5705.011.patch) > [YARN-3368] Add support for Timeline V2 to new web UI > - > > Key: YARN-5705 > URL: https://issues.apache.org/jira/browse/YARN-5705 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sunil G >Assignee: Akhil PB > Attachments: YARN-5705.001.patch, YARN-5705.002.patch, > YARN-5705.003.patch, YARN-5705.004.patch, YARN-5705.005.patch, > YARN-5705.006.patch, YARN-5705.007.patch, YARN-5705.008.patch, > YARN-5705.009.patch, YARN-5705.010.patch, YARN-5705.011.patch > > > Integrate timeline v2 to YARN-3368. This is a clone JIRA for YARN-4097 -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5705) [YARN-3368] Add support for Timeline V2 to new web UI
[ https://issues.apache.org/jira/browse/YARN-5705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB updated YARN-5705: --- Attachment: YARN-5705.011.patch > [YARN-3368] Add support for Timeline V2 to new web UI > - > > Key: YARN-5705 > URL: https://issues.apache.org/jira/browse/YARN-5705 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sunil G >Assignee: Akhil PB > Attachments: YARN-5705.001.patch, YARN-5705.002.patch, > YARN-5705.003.patch, YARN-5705.004.patch, YARN-5705.005.patch, > YARN-5705.006.patch, YARN-5705.007.patch, YARN-5705.008.patch, > YARN-5705.009.patch, YARN-5705.010.patch, YARN-5705.011.patch > > > Integrate timeline v2 to YARN-3368. This is a clone JIRA for YARN-4097 -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-2995) Enhance UI to show cluster resource utilization of various container types
[ https://issues.apache.org/jira/browse/YARN-2995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Konstantinos Karanasos updated YARN-2995: - Attachment: YARN-2995.002.patch Uploading new version of the patch. Rebasing against trunk, addressing [~asuresh]'s comments, fixing existing test cases, adding new test cases. > Enhance UI to show cluster resource utilization of various container types > -- > > Key: YARN-2995 > URL: https://issues.apache.org/jira/browse/YARN-2995 > Project: Hadoop YARN > Issue Type: Sub-task > Components: resourcemanager >Reporter: Sriram Rao >Assignee: Konstantinos Karanasos > Attachments: YARN-2995.001.patch, YARN-2995.002.patch > > > This JIRA proposes to extend the Resource manager UI to show how cluster > resources are being used to run *guaranteed start* and *queueable* > containers. For example, a graph that shows over time, the fraction of > running containers that are *guaranteed start* and the fraction of running > containers that are *queueable*. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610444#comment-15610444 ] Wangda Tan commented on YARN-2009: -- Hi [~eepayne], Really appreciate your help on testing and give feedbacks to this feature from the beginning. We have exactly same requirement: user share queues with different app priority. I was not trying to declare this feature is completed -- I don't think user should use it before we fix the user limit issues you found. My suggestion is: commit existing patch, at least it works for single user and (I think) it doesn't have fundamental issue. We can fix it in the next patch with highest priority. Does this sound OK to you? I just feel reviewing changes of this patch is too painful, every time I felt I'm reviewing a new patch because it's hard to get which place changed :) I think the next patch should be incremental and much easier to be reviewed. Please let me know your thoughts, thanks. > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610374#comment-15610374 ] Bibin A Chundatt commented on YARN-5773: IIUC SchedulerApplicationAttempt#isRecovering is set only in following case .App is in ACCEPTED state i am not sure we will get isRecovery=true {code} // We will replay the final attempt only if last attempt is in final // state but application is not in final state. if (rmApp.getCurrentAppAttempt() == appAttempt && !RMAppImpl.isAppInFinalState(rmApp)) { // Add the previous finished attempt to scheduler synchronously so // that scheduler knows the previous attempt. appAttempt.scheduler.handle(new AppAttemptAddedSchedulerEvent( appAttempt.getAppAttemptId(), false, true)); (new BaseFinalTransition(appAttempt.recoveredFinalState)).transition( appAttempt, event); } {code} {quote} should we update AM diagnostics if we return right from beginning of activateApplications {quote} Cluster resource UI is self explanatory, so required be add ?? {quote} Also, in the patch LOG.debug statements should be guarded with LOG.isDebugEnabled check {quote} Will update the same in next patch > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610265#comment-15610265 ] Hadoop QA commented on YARN-5783: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 57s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 24s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 47s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 20s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 5s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 29s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 29s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 18s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 2 new + 1 unchanged - 0 fixed = 3 total (was 1) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 4s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 35m 53s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 56m 39s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-5783 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835436/yarn-5783.YARN-4752.1.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 3b355260ab84 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | YARN-4752 / 5ad5085 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-YARN-Build/13529/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13529/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13529/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Unit tests to verify the identification of starved applications >
[jira] [Commented] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610221#comment-15610221 ] Hadoop QA commented on YARN-5783: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 19s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 10m 22s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 39s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 21s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 39s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 19s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 58s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 23s{color} | {color:green} YARN-4752 passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 29s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 29s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 17s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 3 new + 0 unchanged - 0 fixed = 3 total (was 0) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 35s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 15s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 1s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 39m 15s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 16s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 58m 14s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-5783 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835434/yarn-5783.YARN-4752.1.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 1a3fc1ff28c1 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | YARN-4752 / 5ad5085 | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-YARN-Build/13527/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13527/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13527/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Unit tests to verify the identification of starved applications > -
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610204#comment-15610204 ] Hadoop QA commented on YARN-5767: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 15s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 40s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 27s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 51s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 16s{color} | {color:green} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager: The patch generated 0 new + 159 unchanged - 27 fixed = 159 total (was 186) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 12s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 15m 29s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 16s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 30m 41s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Issue | YARN-5767 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835439/YARN-5767-trunk-v4.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 26badfc5aa93 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 22ff0ef | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13528/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13528/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versi
[jira] [Commented] (YARN-5690) Integrate native services modules into maven build
[ https://issues.apache.org/jira/browse/YARN-5690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610196#comment-15610196 ] Gour Saha commented on YARN-5690: - [~billie.rinaldi] +1 for the 003 patch. It looks good to me. > Integrate native services modules into maven build > -- > > Key: YARN-5690 > URL: https://issues.apache.org/jira/browse/YARN-5690 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Billie Rinaldi >Assignee: Billie Rinaldi > Attachments: YARN-5690-yarn-native-services.001.patch, > YARN-5690-yarn-native-services.002.patch, > YARN-5690-yarn-native-services.003.patch > > > The yarn dist assembly should include jars for the new modules as well as > their new dependencies. We may want to create new lib directories in the > tarball for the dependencies of the slider-core and services API modules, to > avoid adding these dependencies into the general YARN classpath. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610147#comment-15610147 ] Chris Trezzo commented on YARN-5767: Two notes: # Instead of removing {{LocalCacheCleanerStats.getUserDelSizes()}} I made it return an unmodifiable map. That way, users of the class still have access to the data (outside of the toString method) and it is still protected. # I wound up removing both {{LRUComparator.equals}} and {{LRUComparator.hashCode}}. I figure we don't need to override them since the methods were just using the default implementation anyways. My intention is to file a followup jira that adds metrics that expose the statistics from {{LocalCacheCleanerStats}}. > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch, YARN-5767-trunk-v4.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the target cache size and the public cache is > empty. > # The cache size is larger than the target cache size and everything in the > public cache is being used by a running container. > For clusters that primarily use the public cache (i.e. make use of the shared > cache), this means that the most commonly used resources can be deleted > before old resources in the private cache. Furthermore, the private cache can > continue to grow over time causing more and more churn in the public cache. > Additionally, the same problem exists within the private cache. Since > resources are added to the retention set on a user by user basis, resources > will get cleaned up one user at a time in the order that privateRsrc.values() > returns the LocalResourcesTracker. So if user1 has 10MB in their cache and > user2 has 100MB in their cache and the target size of the cache is 50MB, > user1 could potentially have their entire cache removed before anything is > deleted from the user2 cache. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Trezzo updated YARN-5767: --- Attachment: YARN-5767-trunk-v4.patch Attached is a v4 patch for trunk addressing all comments from the reviews. Thanks! > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch, YARN-5767-trunk-v4.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the target cache size and the public cache is > empty. > # The cache size is larger than the target cache size and everything in the > public cache is being used by a running container. > For clusters that primarily use the public cache (i.e. make use of the shared > cache), this means that the most commonly used resources can be deleted > before old resources in the private cache. Furthermore, the private cache can > continue to grow over time causing more and more churn in the public cache. > Additionally, the same problem exists within the private cache. Since > resources are added to the retention set on a user by user basis, resources > will get cleaned up one user at a time in the order that privateRsrc.values() > returns the LocalResourcesTracker. So if user1 has 10MB in their cache and > user2 has 100MB in their cache and the target size of the cache is 50MB, > user1 could potentially have their entire cache removed before anything is > deleted from the user2 cache. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated YARN-5783: --- Attachment: yarn-5783.YARN-4752.1.patch > Unit tests to verify the identification of starved applications > --- > > Key: YARN-5783 > URL: https://issues.apache.org/jira/browse/YARN-5783 > Project: Hadoop YARN > Issue Type: Sub-task > Components: fairscheduler >Affects Versions: 2.8.0 >Reporter: Karthik Kambatla >Assignee: Karthik Kambatla > Attachments: yarn-5783.YARN-4752.1.patch, yarn-5783.YARN-4752.1.patch > > > JIRA to track unit tests to verify the identification of starved > applications. An application should be marked starved only when: > # Cluster allocation is over the configured threshold for preemption. > # Preemption is enabled for a queue and any of the following: > ## The queue is under its minshare for longer than minsharePreemptionTimeout > ## One of the queue’s applications is under its fairshare for longer than > fairsharePreemptionTimeout. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated YARN-5783: --- Attachment: yarn-5783.YARN-4752.1.patch > Unit tests to verify the identification of starved applications > --- > > Key: YARN-5783 > URL: https://issues.apache.org/jira/browse/YARN-5783 > Project: Hadoop YARN > Issue Type: Sub-task > Components: fairscheduler >Affects Versions: 2.8.0 >Reporter: Karthik Kambatla >Assignee: Karthik Kambatla > Attachments: yarn-5783.YARN-4752.1.patch > > > JIRA to track unit tests to verify the identification of starved > applications. An application should be marked starved only when: > # Cluster allocation is over the configured threshold for preemption. > # Preemption is enabled for a queue and any of the following: > ## The queue is under its minshare for longer than minsharePreemptionTimeout > ## One of the queue’s applications is under its fairshare for longer than > fairsharePreemptionTimeout. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated YARN-5783: --- Attachment: (was: yarn-5783.YARN-4752.1.patch) > Unit tests to verify the identification of starved applications > --- > > Key: YARN-5783 > URL: https://issues.apache.org/jira/browse/YARN-5783 > Project: Hadoop YARN > Issue Type: Sub-task > Components: fairscheduler >Affects Versions: 2.8.0 >Reporter: Karthik Kambatla >Assignee: Karthik Kambatla > > JIRA to track unit tests to verify the identification of starved > applications. An application should be marked starved only when: > # Cluster allocation is over the configured threshold for preemption. > # Preemption is enabled for a queue and any of the following: > ## The queue is under its minshare for longer than minsharePreemptionTimeout > ## One of the queue’s applications is under its fairshare for longer than > fairsharePreemptionTimeout. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5783) Unit tests to verify the identification of starved applications
[ https://issues.apache.org/jira/browse/YARN-5783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated YARN-5783: --- Attachment: yarn-5783.YARN-4752.1.patch Unit test patch that verifies the starvation calculations. > Unit tests to verify the identification of starved applications > --- > > Key: YARN-5783 > URL: https://issues.apache.org/jira/browse/YARN-5783 > Project: Hadoop YARN > Issue Type: Sub-task > Components: fairscheduler >Affects Versions: 2.8.0 >Reporter: Karthik Kambatla >Assignee: Karthik Kambatla > Attachments: yarn-5783.YARN-4752.1.patch > > > JIRA to track unit tests to verify the identification of starved > applications. An application should be marked starved only when: > # Cluster allocation is over the configured threshold for preemption. > # Preemption is enabled for a queue and any of the following: > ## The queue is under its minshare for longer than minsharePreemptionTimeout > ## One of the queue’s applications is under its fairshare for longer than > fairsharePreemptionTimeout. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15610050#comment-15610050 ] Sangjin Lee commented on YARN-5767: --- Thanks for the patch [~ctrezzo]! I agree with the above feedback from Jason and Mikios. Just a couple of more minor comments. 1. Let's make {{LRUComparator}} a private static class. This doesn't need to be visible outside {{LocalCacheCleaner}}. 2. Can we remove {{LocalCacheCleanerStats.getUserDelSizes()}}? This publishes the internal map and breaks the encapsulation. It doesn't appear to be used by any code. 3. Please use a generated {{serialVersionUID}} value instead of a hard-coded value of 1. > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the target cache size and the public cache is > empty. > # The cache size is larger than the target cache size and everything in the > public cache is being used by a running container. > For clusters that primarily use the public cache (i.e. make use of the shared > cache), this means that the most commonly used resources can be deleted > before old resources in the private cache. Furthermore, the private cache can > continue to grow over time causing more and more churn in the public cache. > Additionally, the same problem exists within the private cache. Since > resources are added to the retention set on a user by user basis, resources > will get cleaned up one user at a time in the order that privateRsrc.values() > returns the LocalResourcesTracker. So if user1 has 10MB in their cache and > user2 has 100MB in their cache and the target size of the cache is 50MB, > user1 could potentially have their entire cache removed before anything is > deleted from the user2 cache. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4734) Merge branch:YARN-3368 to trunk
[ https://issues.apache.org/jira/browse/YARN-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609933#comment-15609933 ] Allen Wittenauer commented on YARN-4734: Confirmed -17 no longer applies cleanly. > Merge branch:YARN-3368 to trunk > --- > > Key: YARN-4734 > URL: https://issues.apache.org/jira/browse/YARN-4734 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-4734.1.patch, YARN-4734.10-NOT_READY.patch, > YARN-4734.11-NOT_READY.patch, YARN-4734.12-NOT_READY.patch, > YARN-4734.13.patch, YARN-4734.14.patch, YARN-4734.15.patch, > YARN-4734.16.patch, YARN-4734.17.patch, YARN-4734.2.patch, YARN-4734.3.patch, > YARN-4734.4.patch, YARN-4734.5.patch, YARN-4734.6.patch, YARN-4734.7.patch, > YARN-4734.8.patch, YARN-4734.9-NOT_READY.patch > > > YARN-2928 branch is planned to merge back to trunk shortly, it depends on > changes of YARN-3368. This JIRA is to track the merging task. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4734) Merge branch:YARN-3368 to trunk
[ https://issues.apache.org/jira/browse/YARN-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609878#comment-15609878 ] Allen Wittenauer commented on YARN-4734: I moved the YARN precommit bits to point to Yetus master in order to get this patch run with it under Jenkins. I guess trunk has changed too much for it to apply cleanly. :( > Merge branch:YARN-3368 to trunk > --- > > Key: YARN-4734 > URL: https://issues.apache.org/jira/browse/YARN-4734 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-4734.1.patch, YARN-4734.10-NOT_READY.patch, > YARN-4734.11-NOT_READY.patch, YARN-4734.12-NOT_READY.patch, > YARN-4734.13.patch, YARN-4734.14.patch, YARN-4734.15.patch, > YARN-4734.16.patch, YARN-4734.17.patch, YARN-4734.2.patch, YARN-4734.3.patch, > YARN-4734.4.patch, YARN-4734.5.patch, YARN-4734.6.patch, YARN-4734.7.patch, > YARN-4734.8.patch, YARN-4734.9-NOT_READY.patch > > > YARN-2928 branch is planned to merge back to trunk shortly, it depends on > changes of YARN-3368. This JIRA is to track the merging task. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4734) Merge branch:YARN-3368 to trunk
[ https://issues.apache.org/jira/browse/YARN-4734?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609868#comment-15609868 ] Hadoop QA commented on YARN-4734: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 6s{color} | {color:red} YARN-4734 does not apply to trunk. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | YARN-4734 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835186/YARN-4734.17.patch | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13526/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Merge branch:YARN-3368 to trunk > --- > > Key: YARN-4734 > URL: https://issues.apache.org/jira/browse/YARN-4734 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-4734.1.patch, YARN-4734.10-NOT_READY.patch, > YARN-4734.11-NOT_READY.patch, YARN-4734.12-NOT_READY.patch, > YARN-4734.13.patch, YARN-4734.14.patch, YARN-4734.15.patch, > YARN-4734.16.patch, YARN-4734.17.patch, YARN-4734.2.patch, YARN-4734.3.patch, > YARN-4734.4.patch, YARN-4734.5.patch, YARN-4734.6.patch, YARN-4734.7.patch, > YARN-4734.8.patch, YARN-4734.9-NOT_READY.patch > > > YARN-2928 branch is planned to merge back to trunk shortly, it depends on > changes of YARN-3368. This JIRA is to track the merging task. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609848#comment-15609848 ] Eric Payne commented on YARN-2009: -- bq. there are cases where unnecessary preemption still happens. I have a multi-tenant use case where 1 or 2 intra-queue preemptions occur unnecessarily _every cycle_. bq. many of potential computation mistakes should be able to be corrected automatically by several rounds This one will not. It happens every round. I can provide the details of my manual test if you would like. I haven't found a way yet to change the configs to make this stop happening. If you have suggestions, please let me know. Many of our queues are used by multiple users, and if a high priority app can cause apps from other users to be unnecessarily preempted every preemption cycle, that will not be acceptable by our users. bq. Let's focus all the user-limit related issues in the next patch. The patch works well if users don't share queues. However, my concern is that if users share queues AND they want to use this feature to manage priorities between apps of the same user, this will cause problems with other users in the queue. If we put this in as it is, we are stating that this feature must not be turned on if any preemptable queue in the cluster is multi-tenant. Is that what we want? > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609727#comment-15609727 ] Jason Lowe commented on YARN-5767: -- Ah, I missed the findbugs connection with respect to comparator serialization. In that case I'm fine with leaving that in, since the alternative of adding a findbugs exception is about as much work. No need to do read/write object methods. > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the target cache size and the public cache is > empty. > # The cache size is larger than the target cache size and everything in the > public cache is being used by a running container. > For clusters that primarily use the public cache (i.e. make use of the shared > cache), this means that the most commonly used resources can be deleted > before old resources in the private cache. Furthermore, the private cache can > continue to grow over time causing more and more churn in the public cache. > Additionally, the same problem exists within the private cache. Since > resources are added to the retention set on a user by user basis, resources > will get cleaned up one user at a time in the order that privateRsrc.values() > returns the LocalResourcesTracker. So if user1 has 10MB in their cache and > user2 has 100MB in their cache and the target size of the cache is 50MB, > user1 could potentially have their entire cache removed before anything is > deleted from the user2 cache. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609674#comment-15609674 ] Chris Trezzo commented on YARN-5767: Thanks for the reviews [~miklos.szeg...@cloudera.com] and [~jlowe]! bq. You might want to use resources.getLocalRsrc().containsKey here just like in testPositiveRefCount. [~miklos.szeg...@cloudera.com] Ack, I will add the check to testLRUAcrossTrackers. bq. Why does LRUComparator implement Serializable? It doesn't really do it properly and appears to be unnecessary. I agree that it is unnecessary since we never serialize LRUComparator or the map that we add it to. This is a result of the old code and our findbugs check. There were a couple of modifications I made to the LRUComparator class to satisfy findbugs: # The comparator interface forces you to implement the equals method (how the original code was implemented), but does not force you to add a hashcode implementation. Findbugs now flags this. The original implementation of the equals method was implemented incorrectly (compares object references), but I left it because we never compare comparators in the code. To appease findbugs, I added an identity hashcode implementation that paired with the original implementation of the equals method. In hindsight, maybe I should just fix the original implementation of the equals method and add a correct hashcode implementation as well. [~jlowe], it seems a little unnecessary because our code never uses it, so I would be curious to hear your thoughts. # Findbugs flags the original LRUComparator for implementing the comparator interface, but not implementing Serializable. The argument is that if the comparator is not serializable and it is added to a map, then the map is not serializable. The LRUComparator has no member variables, so I figured all I had to do was declare it serializable and add a serial version number. [~jlowe], should I add default write/read object methods as well? Thanks! bq. Is the concern that someone will create a long-term cache cleaner object that will hold onto these LocalizedResource objects? If so then it would be simpler and faster to just clear the entire resourceMap at the end of the method, since we don't really support reusing a cleaner object. Maybe? This was just copied from the original ResourceRetentionSet class. I can delete the remove call and clear the entire resourceMap at the end of {{cleanCache}}. I will also address the other comments. Thanks again for the reviews! > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609623#comment-15609623 ] Wangda Tan commented on YARN-2009: -- I suggest to commit this patch first and fix other possible problems separately. This patch is already very complex and dense, review modified logics is very painful. Besides user-limit related computations, I don't think it has very critical issues, many of potential computation mistakes should be able to be corrected automatically by several rounds of scheduling/preemption logics (Just like example in the above comment). Let's focus all the user-limit related issues in the next patch. Thoughts? > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609597#comment-15609597 ] Wangda Tan commented on YARN-2009: -- [~eepayne], [~sunilg]. I'm not sure if we need to consider free resource for a queue. There're two scenarios we need to cover 1) One queue's used < configured, and cluster is full 2) One queue's used < configured, and cluster has enough free resource For 1), we should do something for the queue to make sure it can balance, if inter-queue preemption happens at the same time, less intra-queue preemption happens For 2), it is possible that we mark some resource to preemption candidate, but before these resource preempted, the queue will get more resource and previously added candidates will be cancelled. And also, the whole point of intra-queue preemption is to make sure queue can balance itself. We have configs for intra-queue preemption minimum-threshold/max-allowable-limit to make sure we will 1) Preempt queue's resource when it is not beyond a threshold 2) Do not preempt too much resource In addition, inter-queue preemption happens first before intra-queue preemption. With all the above safeguards, existing logics looks safe since we won't kill containers immediately after they get selected. I'm a little concern that considering the (configured - used) resource of queue could cause sometimes intra-queue preemption cannot be triggered. > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5771) Provide option to send env to be whitelisted in ContainerLaunchContext
[ https://issues.apache.org/jira/browse/YARN-5771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609586#comment-15609586 ] Daniel Templeton commented on YARN-5771: #2 is required to keep this approach from reopening the security hole. Even with that, though, it makes me a little uncomfortable. I'd prefer security to be inclusive ("allow these things") rather than exclusive ("don't allow these things") because it's easy to forget to add something to the black list. What about the approach outlined in [https://hadoop.apache.org/docs/r3.0.0-alpha1/hadoop-mapreduce-client/hadoop-mapreduce-client-core/DistributedCacheDeploy.html]? It solves the issue handily. You can also just add {noformat} mapreduce.admin.user.env HADOOP_MAPRED_HOME=$HADOOP_MAPRED_HOME yarn.app.mapreduce.am.env HADOOP_MAPRED_HOME=$HADOOP_MAPRED_HOME {noformat} to the {{mapred-site.xml}} file to resolve the issue in many cases. In my opinion the problem is not so much that anything is broken or needs to be fixed, but that we need to do a better job of documenting the options for configuring a cluster in the post-env-var-leak-fix world. > Provide option to send env to be whitelisted in ContainerLaunchContext > --- > > Key: YARN-5771 > URL: https://issues.apache.org/jira/browse/YARN-5771 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: container-whitelist-env-wip.patch > > > As per current implementation ENV to be white listed for container launch is > are configured as part of {{yarn.nodemanager.env-whitelist}} > Specific to container we cannot specify additional ENV properties to be > whitelisted. As part of this jira we are providing an option to provide > additional whitelist ENV. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5716) Add global scheduler interface definition and update CapacityScheduler to use it.
[ https://issues.apache.org/jira/browse/YARN-5716?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609567#comment-15609567 ] Sunil G commented on YARN-5716: --- Thanks [~leftnoteasy] for the clarifications. Makes sense. New patch looks fine for me too. > Add global scheduler interface definition and update CapacityScheduler to use > it. > - > > Key: YARN-5716 > URL: https://issues.apache.org/jira/browse/YARN-5716 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-5716.001.patch, YARN-5716.002.patch, > YARN-5716.003.patch, YARN-5716.004.patch, YARN-5716.005.patch, > YARN-5716.006.patch, YARN-5716.007.patch, YARN-5716.008.patch, > YARN-5716.009.patch > > > Target of this JIRA: > - Definition of interfaces / objects which will be used by global scheduling, > this will be shared by different schedulers. > - Modify CapacityScheduler to use it. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609409#comment-15609409 ] Sunil G commented on YARN-2009: --- I understand the reason for this issue. Queue's free resource were not considered in the algorithm while calculating idealAssigned. So current code will work in cases where queue is full or overflowed. If queue is under-utilized, we may do over preemption. I ll share a patch short while. > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5433) Audit dependencies for Category-X
[ https://issues.apache.org/jira/browse/YARN-5433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609349#comment-15609349 ] Hudson commented on YARN-5433: -- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10690 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10690/]) YARN-5433. Audit dependencies for Category-X. Contributed by Sangjin (sjlee: rev f511cc89b66997e496f630bdd299d3068d43fd31) * (edit) LICENSE.txt * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/pom.xml * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase-tests/pom.xml > Audit dependencies for Category-X > - > > Key: YARN-5433 > URL: https://issues.apache.org/jira/browse/YARN-5433 > Project: Hadoop YARN > Issue Type: Bug > Components: timelineserver >Affects Versions: 3.0.0-alpha1 >Reporter: Sean Busbey >Assignee: Sangjin Lee >Priority: Blocker > Fix For: 3.0.0-alpha2 > > Attachments: YARN-5433.01.patch, YARN-5433.02.patch, > YARN-5433.03.patch, YARN-5433.04.patch > > > Recently phoenix has found some category-x dependencies in their build > (PHOENIX-3084, PHOENIX-3091), which also showed some problems in HBase > (HBASE-16260). > Since the Timeline Server work brought in both of these as dependencies, we > should make sure we don't have any cat-x dependencies either. From what I've > seen in those projects, our choice of HBase version shouldn't be impacted but > our Phoenix one is. > Greping our current dependency list for the timeline server component shows > some LGPL: > {code} > ... > [INFO]net.sourceforge.findbugs:annotations:jar:1.3.2:compile > ... > {code} > I haven't checked the rest of the dependencies that have changed since > HADOOP-12893 went in, so ATM I've filed this against YARN since that's where > this one example came in. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-3525) Rename fair scheduler properties increment-allocation-mb and increment-allocation-vcores
[ https://issues.apache.org/jira/browse/YARN-3525?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-3525: --- Assignee: (was: Bibin A Chundatt) > Rename fair scheduler properties increment-allocation-mb and > increment-allocation-vcores > > > Key: YARN-3525 > URL: https://issues.apache.org/jira/browse/YARN-3525 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Priority: Minor > > Rename below two properties since only used by fair scheduler > {color:blue}yarn.scheduler.increment-allocation-mb{color} to > {color:red}yarn.scheduler.fair.increment-allocation-mb{color} > {color:blue}yarn.scheduler.increment-allocation-vcores{color} to > {color:red}yarn.scheduler.fair.increment-allocation-vcores{color} > All other properties only for fair scheduler are using {color:red} > yarn.scheduler.fair{color} prefix . -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-4498) Application level node labels stats to be available in UI/REST/CLI
[ https://issues.apache.org/jira/browse/YARN-4498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-4498: --- Attachment: apps.xml Attaching sample xml file for application rest API For each partition the usage stats will be shown {noformat} labelx 153611536100 {noformat} > Application level node labels stats to be available in UI/REST/CLI > -- > > Key: YARN-4498 > URL: https://issues.apache.org/jira/browse/YARN-4498 > Project: Hadoop YARN > Issue Type: Sub-task > Components: api, client, resourcemanager >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: 0001-YARN-4498.patch, YARN-4498.0002.patch, apps.xml > > > Currently nodelabel stats per application is not available through > REST/CLI/UI like currently used labels by all live containers, total stats of > containers per label for app etc.. > Will split into multiple task if approach is fine. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5433) Audit dependencies for Category-X
[ https://issues.apache.org/jira/browse/YARN-5433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609217#comment-15609217 ] Hadoop QA commented on YARN-5433: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 12s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s {color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 35s {color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 50s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 48s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 9m 14s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 2s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 4m 17s {color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 14s {color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 5s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 48s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 48s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 9m 7s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 0s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 3s {color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 4m 16s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 82m 8s {color} | {color:red} root in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 23s {color} | {color:red} The patch generated 2 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 141m 38s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.nodemanager.containermanager.queuing.TestQueuingContainerManager | | Timed out junit tests | org.apache.hadoop.hdfs.server.datanode.TestDataNodeRollingUpgrade | | | org.apache.hadoop.hdfs.server.diskbalancer.command.TestDiskBalancerCommand | | | org.apache.hadoop.hdfs.server.diskbalancer.TestDiskBalancerRPC | | | org.apache.hadoop.hdfs.server.datanode.TestDirectoryScanner | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835368/YARN-5433.04.patch | | JIRA Issue | YARN-5433 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit xml | | uname | Linux fff145610caa 3.13.0-93-generic #140-Ubuntu SMP Mon Jul 18 21:21:05 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 9cad3e2 | | Default Java | 1.8.0_101 | | unit | https://builds.apache.org/job/PreCommit-YARN-Build/13523/artifact/patchprocess/patch-unit-root.txt | | unit test logs | https://builds.apache.org/job/PreCommit-YARN-Build/13523/artifact/patchprocess/patch-unit-root.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13523/testReport/ | | asflicense | https://builds.apache.org/job/PreCommit-YARN-Build/13523/artifact/patchprocess/patch-asflicense-problems.txt | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice hadoop-yarn-project/hadoop-yarn/hadoop-ya
[jira] [Updated] (YARN-4498) Application level node labels stats to be available in UI/REST/CLI
[ https://issues.apache.org/jira/browse/YARN-4498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-4498: --- Attachment: YARN-4498.0002.patch Attaching updated patch after adding partition level usage to rest API > Application level node labels stats to be available in UI/REST/CLI > -- > > Key: YARN-4498 > URL: https://issues.apache.org/jira/browse/YARN-4498 > Project: Hadoop YARN > Issue Type: Sub-task > Components: api, client, resourcemanager >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: 0001-YARN-4498.patch, YARN-4498.0002.patch > > > Currently nodelabel stats per application is not available through > REST/CLI/UI like currently used labels by all live containers, total stats of > containers per label for app etc.. > Will split into multiple task if approach is fine. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609208#comment-15609208 ] Varun Saxena commented on YARN-5773: bq. Now the question here is to invoke activateApplications after scheduler recovery is done. Do we need to activateApplications ? Applications would be activated when node re-registers. Right ? Coming to code in the patch, checking against cluster resource being 0 should be fine but resources may not be 0 only on recovery so should we update AM diagnostics if we return right from beginning of activateApplications ? Also a log should be added. One more thing, the indication of whether attempt is recovering or not can be received from SchedulerApplicationAttempt#isRecovering which is filled from AppAttemptAddedSchedulerEvent. This can be used as well to differentiate between 0 resources because app is recovering or no nodes registered. Also, in the patch LOG.debug statements should be guarded with LOG.isDebugEnabled check > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-5783) Unit tests to verify the identification of starved applications
Karthik Kambatla created YARN-5783: -- Summary: Unit tests to verify the identification of starved applications Key: YARN-5783 URL: https://issues.apache.org/jira/browse/YARN-5783 Project: Hadoop YARN Issue Type: Sub-task Components: fairscheduler Affects Versions: 2.8.0 Reporter: Karthik Kambatla Assignee: Karthik Kambatla JIRA to track unit tests to verify the identification of starved applications. An application should be marked starved only when: # Cluster allocation is over the configured threshold for preemption. # Preemption is enabled for a queue and any of the following: ## The queue is under its minshare for longer than minsharePreemptionTimeout ## One of the queue’s applications is under its fairshare for longer than fairsharePreemptionTimeout. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609101#comment-15609101 ] Sunil G commented on YARN-2009: --- Thanks [~eepayne] for the update. I will also investigate with few more scenarios w.r.t user-limit and priority together. For the testcase, I think there is a stubbing issue in mockLeafQueue. I can make change for the same. Meanwhile I will wait for some more manual tests from my end, and will try to share a report here. Meantime pls let me know if you run into any issues. > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-2009) Priority support for preemption in ProportionalCapacityPreemptionPolicy
[ https://issues.apache.org/jira/browse/YARN-2009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15609030#comment-15609030 ] Eric Payne commented on YARN-2009: -- [~sunilg], I am still analyzing the code, but I do want to point out one concern that I have. {{TestProportionalCapacityPreemptionPolicyIntraQueue#testPreemptionWithTwoUsers}}: In this test, the user limit resource is 30, and {{app3}} has 40 resources. So, at most, I would expect only 10 resources to be preempted from {{app3}}, which would bring {{app3}} down to 30 resources to match the ULR. Instead, 25 resources were preempted. I feel that {{FifoIntraQueuePreemptionPlugin#calculateIdealAssignedResourcePerApp}} should not allow {{tmpApp.idealAssigned}} to be lower than the ULR. One more thing I noticed during my manual testing is that although the unnecessary preemption is a lot less, there are cases where it still happens. I haven't drilled down yet. > Priority support for preemption in ProportionalCapacityPreemptionPolicy > --- > > Key: YARN-2009 > URL: https://issues.apache.org/jira/browse/YARN-2009 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler >Reporter: Devaraj K >Assignee: Sunil G > Attachments: YARN-2009.0001.patch, YARN-2009.0002.patch, > YARN-2009.0003.patch, YARN-2009.0004.patch, YARN-2009.0005.patch, > YARN-2009.0006.patch, YARN-2009.0007.patch, YARN-2009.0008.patch, > YARN-2009.0009.patch, YARN-2009.0010.patch, YARN-2009.0011.patch, > YARN-2009.0012.patch, YARN-2009.0013.patch, YARN-2009.0014.patch, > YARN-2009.0015.patch > > > While preempting containers based on the queue ideal assignment, we may need > to consider preempting the low priority application containers first. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4330) MiniYARNCluster is showing multiple Failed to instantiate default resource calculator warning messages.
[ https://issues.apache.org/jira/browse/YARN-4330?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608994#comment-15608994 ] Varun Saxena commented on YARN-4330: The patch no longer applies cleanly. Give me a day, I will put up a new patch for review. Also after YARN-5662, there is a slight modification required in the patch. > MiniYARNCluster is showing multiple Failed to instantiate default resource > calculator warning messages. > > > Key: YARN-4330 > URL: https://issues.apache.org/jira/browse/YARN-4330 > Project: Hadoop YARN > Issue Type: Bug > Components: test, yarn >Affects Versions: 2.8.0 > Environment: OSX, JUnit >Reporter: Steve Loughran >Assignee: Varun Saxena >Priority: Blocker > Attachments: YARN-4330.01.patch > > > Whenever I try to start a MiniYARNCluster on Branch-2 (commit #0b61cca), I > see multiple stack traces warning me that a resource calculator plugin could > not be created > {code} > (ResourceCalculatorPlugin.java:getResourceCalculatorPlugin(184)) - > java.lang.UnsupportedOperationException: Could not determine OS: Failed to > instantiate default resource calculator. > java.lang.UnsupportedOperationException: Could not determine OS > {code} > This is a minicluster. It doesn't need resource calculation. It certainly > doesn't need test logs being cluttered with even more stack traces which will > only generate false alarms about tests failing. > There needs to be a way to turn this off, and the minicluster should have it > that way by default. > Being ruthless and marking as a blocker, because its a fairly major > regression for anyone testing with the minicluster. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5752) TestLocalResourcesTrackerImpl#testLocalResourceCache times out
[ https://issues.apache.org/jira/browse/YARN-5752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608976#comment-15608976 ] Miklos Szegedi commented on YARN-5752: -- +1 non-binding. The patch looks good to me. Thank you, [~ebadger]! > TestLocalResourcesTrackerImpl#testLocalResourceCache times out > -- > > Key: YARN-5752 > URL: https://issues.apache.org/jira/browse/YARN-5752 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Eric Badger >Assignee: Eric Badger > Attachments: YARN-5752.001.patch, YARN-5752.002.patch > > > {noformat} > java.lang.Exception: test timed out after 1000 milliseconds > at sun.nio.cs.StreamEncoder.write(StreamEncoder.java:133) > at java.io.OutputStreamWriter.write(OutputStreamWriter.java:220) > at java.io.Writer.write(Writer.java:157) > at org.apache.log4j.helpers.QuietWriter.write(QuietWriter.java:48) > at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:310) > at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) > at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) > at > org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) > at org.apache.log4j.Category.callAppenders(Category.java:206) > at org.apache.log4j.Category.forcedLog(Category.java:391) > at org.apache.log4j.Category.log(Category.java:856) > at > org.apache.commons.logging.impl.Log4JLogger.info(Log4JLogger.java:176) > at > org.apache.hadoop.yarn.event.AsyncDispatcher.register(AsyncDispatcher.java:209) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.TestLocalResourcesTrackerImpl.testLocalResourceCache(TestLocalResourcesTrackerImpl.java:258) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5779) [YARN-3368] Document limits/notes of the new YARN UI
[ https://issues.apache.org/jira/browse/YARN-5779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608926#comment-15608926 ] Sunil G commented on YARN-5779: --- +1. Committing > [YARN-3368] Document limits/notes of the new YARN UI > > > Key: YARN-5779 > URL: https://issues.apache.org/jira/browse/YARN-5779 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-5779-YARN-3368.01.patch, > YARN-5779-YARN-3368.02.patch > > > For example, we don't make sure it's able to run on security enabled > environment. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5779) [YARN-3368] Document limits/notes of the new YARN UI
[ https://issues.apache.org/jira/browse/YARN-5779?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608918#comment-15608918 ] Hadoop QA commented on YARN-5779: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 3m 43s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 11m 3s {color} | {color:green} YARN-3368 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 18s {color} | {color:green} YARN-3368 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 11s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 17s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 15m 47s {color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:5a4801a | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835373/YARN-5779-YARN-3368.02.patch | | JIRA Issue | YARN-5779 | | Optional Tests | asflicense mvnsite | | uname | Linux 696240431c8c 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | YARN-3368 / 9690f29 | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13524/console | | Powered by | Apache Yetus 0.3.0 http://yetus.apache.org | This message was automatically generated. > [YARN-3368] Document limits/notes of the new YARN UI > > > Key: YARN-5779 > URL: https://issues.apache.org/jira/browse/YARN-5779 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-5779-YARN-3368.01.patch, > YARN-5779-YARN-3368.02.patch > > > For example, we don't make sure it's able to run on security enabled > environment. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5428) Allow for specifying the docker client configuration directory
[ https://issues.apache.org/jira/browse/YARN-5428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608907#comment-15608907 ] Allen Wittenauer commented on YARN-5428: It's already not particularly interested. Mesos and Kubernetes have way too much momentum and mind share for anyone to deal with YARN for general Docker management. > Allow for specifying the docker client configuration directory > -- > > Key: YARN-5428 > URL: https://issues.apache.org/jira/browse/YARN-5428 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn >Reporter: Shane Kumpf >Assignee: Shane Kumpf > Attachments: YARN-5428.001.patch, YARN-5428.002.patch, > YARN-5428.003.patch, YARN-5428.004.patch, > YARN-5428Allowforspecifyingthedockerclientconfigurationdirectory.pdf > > > The docker client allows for specifying a configuration directory that > contains the docker client's configuration. It is common to store "docker > login" credentials in this config, to avoid the need to docker login on each > cluster member. > By default the docker client config is $HOME/.docker/config.json on Linux. > However, this does not work with the current container executor user > switching and it may also be desirable to centralize this configuration > beyond the single user's home directory. > Note that the command line arg is for the configuration directory NOT the > configuration file. > This change will be needed to allow YARN to automatically pull images at > localization time or within container executor. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5767) Fix the order that resources are cleaned up from the local Public/Private caches
[ https://issues.apache.org/jira/browse/YARN-5767?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608897#comment-15608897 ] Jason Lowe commented on YARN-5767: -- Thanks for the patch, Chris! Looks good overall, just some minor nits in addition to the unit test comment above. I don't think the actual resource path logging is necessary since LocalResourcesTrackerImpl already does this in its remove method. If this does get added here, please make it a trace-level log separate from the debug one. LocalCacheCleaner should be a package-private class. It does not need to be public. handleCacheCleanup takes an event that isn't used and should be removed. Looks like this was in the original code as well, but it would be nice to cleanup while we're here. Why does LRUComparator implement Serializable? It doesn't really do it properly and appears to be unnecessary. LocalCacheCleanerStats should take the currentSize as a constructor argument which can be used to initialize cacheSizeBeforeClean. As a bonus, cacheSizeBeforeClean can then be marked final. I'm curious why we're removing the map entry here: {code} if (tracker.remove(resource, delService)) { stats.incDelSize(tracker.getUser(), resource.getSize()); i.remove(); } {code} The removal seems unnecessary. Is the concern that someone will create a long-term cache cleaner object that will hold onto these LocalizedResource objects? If so then it would be simpler and faster to just clear the entire resourceMap at the end of the method, since we don't really support reusing a cleaner object. > Fix the order that resources are cleaned up from the local Public/Private > caches > > > Key: YARN-5767 > URL: https://issues.apache.org/jira/browse/YARN-5767 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 2.6.0, 2.7.0, 3.0.0-alpha1 >Reporter: Chris Trezzo >Assignee: Chris Trezzo > Attachments: YARN-5767-trunk-v1.patch, YARN-5767-trunk-v2.patch, > YARN-5767-trunk-v3.patch > > > If you look at {{ResourceLocalizationService#handleCacheCleanup}}, you can > see that public resources are added to the {{ResourceRetentionSet}} first > followed by private resources: > {code:java} > private void handleCacheCleanup(LocalizationEvent event) { > ResourceRetentionSet retain = > new ResourceRetentionSet(delService, cacheTargetSize); > retain.addResources(publicRsrc); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup (public) " + retain); > } > for (LocalResourcesTracker t : privateRsrc.values()) { > retain.addResources(t); > if (LOG.isDebugEnabled()) { > LOG.debug("Resource cleanup " + t.getUser() + ":" + retain); > } > } > //TODO Check if appRsrcs should also be added to the retention set. > } > {code} > Unfortunately, if we look at {{ResourceRetentionSet#addResources}} we see > that this means public resources are deleted first until the target cache > size is met: > {code:java} > public void addResources(LocalResourcesTracker newTracker) { > for (LocalizedResource resource : newTracker) { > currentSize += resource.getSize(); > if (resource.getRefCount() > 0) { > // always retain resources in use > continue; > } > retain.put(resource, newTracker); > } > for (Iterator> i = > retain.entrySet().iterator(); >currentSize - delSize > targetSize && i.hasNext();) { > Map.Entry rsrc = i.next(); > LocalizedResource resource = rsrc.getKey(); > LocalResourcesTracker tracker = rsrc.getValue(); > if (tracker.remove(resource, delService)) { > delSize += resource.getSize(); > i.remove(); > } > } > } > {code} > The result of this is that resources in the private cache are only deleted in > the cases where: > # The cache size is larger than the target cache size and the public cache is > empty. > # The cache size is larger than the target cache size and everything in the > public cache is being used by a running container. > For clusters that primarily use the public cache (i.e. make use of the shared > cache), this means that the most commonly used resources can be deleted > before old resources in the private cache. Furthermore, the private cache can > continue to grow over time causing more and more churn in the public cache. > Additionally, the same problem exists within the private cache. Since > resources are added to the retention set on a user by user basis, resources > will get cleaned up one user at a time in the order that privateRsrc.values() > returns the LocalResourcesTracker. So if user1 has 10MB in their cache and > user2 has 100MB in their cache and the target size of the cache is 50MB, > user1 could potentially have their en
[jira] [Comment Edited] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607750#comment-15607750 ] Bibin A Chundatt edited comment on YARN-5773 at 10/26/16 4:15 PM: -- [~sunilg] have raise JIRA YARN-5781 for handling optimization cases was (Author: bibinchundatt): [~sunilg] have raise JIRA YARN-5781 for the same. > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5779) [YARN-3368] Document limits/notes of the new YARN UI
[ https://issues.apache.org/jira/browse/YARN-5779?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sunil G updated YARN-5779: -- Attachment: YARN-5779-YARN-3368.02.patch Patch generally looks fine. However a small typo, correcting and uploading new patch > [YARN-3368] Document limits/notes of the new YARN UI > > > Key: YARN-5779 > URL: https://issues.apache.org/jira/browse/YARN-5779 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Wangda Tan >Assignee: Wangda Tan > Attachments: YARN-5779-YARN-3368.01.patch, > YARN-5779-YARN-3368.02.patch > > > For example, we don't make sure it's able to run on security enabled > environment. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5611) Provide an API to update lifetime of an application.
[ https://issues.apache.org/jira/browse/YARN-5611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608854#comment-15608854 ] Hadoop QA commented on YARN-5611: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 9 new or modified test files. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s {color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 42s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 59s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 36s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 31s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 11s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 53s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 35s {color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 15s {color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 2m 3s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 51s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} cc {color} | {color:green} 6m 51s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 51s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} checkstyle {color} | {color:red} 1m 36s {color} | {color:red} root: The patch generated 46 new + 703 unchanged - 4 fixed = 749 total (was 707) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 2m 32s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 12s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} whitespace {color} | {color:red} 0m 0s {color} | {color:red} The patch has 10 line(s) that end in whitespace. Use git apply --whitespace=fix. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 1s {color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 4m 37s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} javadoc {color} | {color:red} 0m 17s {color} | {color:red} hadoop-yarn-project_hadoop-yarn_hadoop-yarn-api generated 2 new + 123 unchanged - 0 fixed = 125 total (was 123) {color} | | {color:red}-1{color} | {color:red} javadoc {color} | {color:red} 0m 21s {color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 26s {color} | {color:green} hadoop-yarn-api in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 2m 39s {color} | {color:green} hadoop-yarn-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 15m 16s {color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 39m 36s {color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 104m 9s {color} | {color:green} hadoop-mapreduce-client-jobclient in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 30s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 209m 34s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.resourcemanager.rm
[jira] [Commented] (YARN-5770) Performance improvement of native-services REST API service
[ https://issues.apache.org/jira/browse/YARN-5770?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608851#comment-15608851 ] Billie Rinaldi commented on YARN-5770: -- +1, the latest patch looks good. > Performance improvement of native-services REST API service > --- > > Key: YARN-5770 > URL: https://issues.apache.org/jira/browse/YARN-5770 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Gour Saha >Assignee: Gour Saha > Fix For: yarn-native-services > > Attachments: YARN-5770-yarn-native-services.003.patch, > YARN-5770-yarn-native-services.004.patch, > YARN-5770-yarn-native-services.phase1.001.patch, > YARN-5770-yarn-native-services.phase1.002.patch > > > Make enhancements and bug-fixes to eliminate frequent full GC of the REST API > Service. Dependent on few Slider fixes like SLIDER-1168 as well. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-5782) Slider version output includes variables buildNumber and hadoop.version
Billie Rinaldi created YARN-5782: Summary: Slider version output includes variables buildNumber and hadoop.version Key: YARN-5782 URL: https://issues.apache.org/jira/browse/YARN-5782 Project: Hadoop YARN Issue Type: Sub-task Reporter: Billie Rinaldi The hadoop.version variable could be switched to project.version, and perhaps buildNumber could be removed since Hadoop does not use the buildnumber plugin. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5433) Audit dependencies for Category-X
[ https://issues.apache.org/jira/browse/YARN-5433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sangjin Lee updated YARN-5433: -- Attachment: YARN-5433.04.patch Posted patch v.4. Removed the exhibit portion of MPL. Thanks [~xiaochen] for checking. I'll wait for a little bit for others to comment, and commit it unless there is an objection. > Audit dependencies for Category-X > - > > Key: YARN-5433 > URL: https://issues.apache.org/jira/browse/YARN-5433 > Project: Hadoop YARN > Issue Type: Bug > Components: timelineserver >Affects Versions: 3.0.0-alpha1 >Reporter: Sean Busbey >Assignee: Sangjin Lee >Priority: Blocker > Attachments: YARN-5433.01.patch, YARN-5433.02.patch, > YARN-5433.03.patch, YARN-5433.04.patch > > > Recently phoenix has found some category-x dependencies in their build > (PHOENIX-3084, PHOENIX-3091), which also showed some problems in HBase > (HBASE-16260). > Since the Timeline Server work brought in both of these as dependencies, we > should make sure we don't have any cat-x dependencies either. From what I've > seen in those projects, our choice of HBase version shouldn't be impacted but > our Phoenix one is. > Greping our current dependency list for the timeline server component shows > some LGPL: > {code} > ... > [INFO]net.sourceforge.findbugs:annotations:jar:1.3.2:compile > ... > {code} > I haven't checked the rest of the dependencies that have changed since > HADOOP-12893 went in, so ATM I've filed this against YARN since that's where > this one example came in. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5752) TestLocalResourcesTrackerImpl#testLocalResourceCache times out
[ https://issues.apache.org/jira/browse/YARN-5752?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Badger updated YARN-5752: -- Attachment: YARN-5752.002.patch [~miklos.szeg...@cloudera.com], yea the 100 second timeout on testHierarchicalLocalCacheDirectories does seem excessive given its consistent sub-1 second runtime. Attaching patch to set timeout to 10 seconds for both. > TestLocalResourcesTrackerImpl#testLocalResourceCache times out > -- > > Key: YARN-5752 > URL: https://issues.apache.org/jira/browse/YARN-5752 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Eric Badger >Assignee: Eric Badger > Attachments: YARN-5752.001.patch, YARN-5752.002.patch > > > {noformat} > java.lang.Exception: test timed out after 1000 milliseconds > at sun.nio.cs.StreamEncoder.write(StreamEncoder.java:133) > at java.io.OutputStreamWriter.write(OutputStreamWriter.java:220) > at java.io.Writer.write(Writer.java:157) > at org.apache.log4j.helpers.QuietWriter.write(QuietWriter.java:48) > at org.apache.log4j.WriterAppender.subAppend(WriterAppender.java:310) > at org.apache.log4j.WriterAppender.append(WriterAppender.java:162) > at org.apache.log4j.AppenderSkeleton.doAppend(AppenderSkeleton.java:251) > at > org.apache.log4j.helpers.AppenderAttachableImpl.appendLoopOnAppenders(AppenderAttachableImpl.java:66) > at org.apache.log4j.Category.callAppenders(Category.java:206) > at org.apache.log4j.Category.forcedLog(Category.java:391) > at org.apache.log4j.Category.log(Category.java:856) > at > org.apache.commons.logging.impl.Log4JLogger.info(Log4JLogger.java:176) > at > org.apache.hadoop.yarn.event.AsyncDispatcher.register(AsyncDispatcher.java:209) > at > org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.TestLocalResourcesTrackerImpl.testLocalResourceCache(TestLocalResourcesTrackerImpl.java:258) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608433#comment-15608433 ] Hadoop QA commented on YARN-5773: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s {color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 46s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 32s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 21s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 38s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 16s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 58s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 30s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 18s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 36s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 3s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 18s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 35m 55s {color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s {color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 50m 23s {color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835100/YARN-5773.003.patch | | JIRA Issue | YARN-5773 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 76552f764153 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 24a83fe | | Default Java | 1.8.0_101 | | findbugs | v3.0.0 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13521/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13521/console | | Powered by | Apache Yetus 0.3.0 http://yetus.apache.org | This message was automatically generated. > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >
[jira] [Updated] (YARN-5705) [YARN-3368] Add support for Timeline V2 to new web UI
[ https://issues.apache.org/jira/browse/YARN-5705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akhil PB updated YARN-5705: --- Attachment: YARN-5705.011.patch > [YARN-3368] Add support for Timeline V2 to new web UI > - > > Key: YARN-5705 > URL: https://issues.apache.org/jira/browse/YARN-5705 > Project: Hadoop YARN > Issue Type: Sub-task >Reporter: Sunil G >Assignee: Akhil PB > Attachments: YARN-5705.001.patch, YARN-5705.002.patch, > YARN-5705.003.patch, YARN-5705.004.patch, YARN-5705.005.patch, > YARN-5705.006.patch, YARN-5705.007.patch, YARN-5705.008.patch, > YARN-5705.009.patch, YARN-5705.010.patch, YARN-5705.011.patch > > > Integrate timeline v2 to YARN-3368. This is a clone JIRA for YARN-4097 -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5611) Provide an API to update lifetime of an application.
[ https://issues.apache.org/jira/browse/YARN-5611?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Rohith Sharma K S updated YARN-5611: Attachment: YARN-5611.0004.patch Attaching the patch for an update API server side implementation. Though patch looks bit heavy, most of the changes are in test. Apart from test modifications, the actual modification is as below. # There is protos added in ApplicationClientProtocol for updateApplicationTimeout. This adds 4 files corresponding request/response with PBImpl. # Similarly, there are protos added AppUpdateAttributes to store in StateStore. This adds 4 new files with PBImpl. The main module interaction point is {{ClientRMService-->RMAppImpl-->RMStateStore}}. The interaction flow is as below. *ClientRMService* : Accept incoming request from client and validate requested data. Trigger an event to RMAppImpl to update app timeout. And wait for update success response. In case of state store failure, roll back updated attributes. *RMAppImpl* : It handles new event APP_UPDATE and update in monitoring service. And trigger an event to update in State Store in order. *RMStateStore* : Store updated app state and set the future object which is passed by clientRMService. Once RMStateStore sets future object, clientRMService return to user as success. Pending task : Currently update response is returning empty. But response should contains updated value. I will update this behavior in next patches. > Provide an API to update lifetime of an application. > > > Key: YARN-5611 > URL: https://issues.apache.org/jira/browse/YARN-5611 > Project: Hadoop YARN > Issue Type: Sub-task > Components: resourcemanager >Reporter: Rohith Sharma K S >Assignee: Rohith Sharma K S > Attachments: 0001-YARN-5611.patch, 0002-YARN-5611.patch, > 0003-YARN-5611.patch, YARN-5611.0004.patch, YARN-5611.v0.patch > > > YARN-4205 monitors an Lifetime of an applications is monitored if required. > Add an client api to update lifetime of an application. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5428) Allow for specifying the docker client configuration directory
[ https://issues.apache.org/jira/browse/YARN-5428?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15608044#comment-15608044 ] Zhankun Tang commented on YARN-5428: Do you mean that the community won't be interested on Docker/YARN in the future? > Allow for specifying the docker client configuration directory > -- > > Key: YARN-5428 > URL: https://issues.apache.org/jira/browse/YARN-5428 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn >Reporter: Shane Kumpf >Assignee: Shane Kumpf > Attachments: YARN-5428.001.patch, YARN-5428.002.patch, > YARN-5428.003.patch, YARN-5428.004.patch, > YARN-5428Allowforspecifyingthedockerclientconfigurationdirectory.pdf > > > The docker client allows for specifying a configuration directory that > contains the docker client's configuration. It is common to store "docker > login" credentials in this config, to avoid the need to docker login on each > cluster member. > By default the docker client config is $HOME/.docker/config.json on Linux. > However, this does not work with the current container executor user > switching and it may also be desirable to centralize this configuration > beyond the single user's home directory. > Note that the command line arg is for the configuration directory NOT the > configuration file. > This change will be needed to allow YARN to automatically pull images at > localization time or within container executor. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5433) Audit dependencies for Category-X
[ https://issues.apache.org/jira/browse/YARN-5433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607868#comment-15607868 ] Hadoop QA commented on YARN-5433: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 18s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s {color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 37s {color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 44s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 55s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 9m 22s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 6s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 4m 22s {color} | {color:green} trunk passed {color} | | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 15s {color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 18s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 6m 58s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 6m 58s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 9m 22s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 1m 2s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 3s {color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 4m 26s {color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 101m 15s {color} | {color:red} root in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 21s {color} | {color:red} The patch generated 2 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 161m 59s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.server.datanode.TestLargeBlockReport | | Timed out junit tests | org.apache.hadoop.hdfs.TestLeaseRecovery2 | | | org.apache.hadoop.hdfs.TestParallelRead | | | org.apache.hadoop.hdfs.TestDFSStripedOutputStreamWithFailure160 | | | org.apache.hadoop.hdfs.TestPread | | | org.apache.hadoop.hdfs.TestBlockMissingException | | | org.apache.hadoop.hdfs.TestDFSClientRetries | | | org.apache.hadoop.hdfs.server.balancer.TestBalancer | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:9560f25 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12835266/YARN-5433.03.patch | | JIRA Issue | YARN-5433 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit xml | | uname | Linux 5438fefa4daf 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 287efff | | Default Java | 1.8.0_101 | | unit | https://builds.apache.org/job/PreCommit-YARN-Build/13518/artifact/patchprocess/patch-unit-root.txt | | unit test logs | https://builds.apache.org/job/PreCommit-YARN-Build/13518/artifact/patchprocess/patch-unit-root.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13518/testReport/ | | asflicense | https://builds.apache.org/job/PreCommit-YARN-Build/13518/artifact/patchprocess/patch-asflicense-problems.txt | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelinese
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607750#comment-15607750 ] Bibin A Chundatt commented on YARN-5773: [~sunilg] have raise JIRA YARN-5781 for the same. > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-5781) Optimize LeafQueue#activateApplication() during AM resource limit check
Bibin A Chundatt created YARN-5781: -- Summary: Optimize LeafQueue#activateApplication() during AM resource limit check Key: YARN-5781 URL: https://issues.apache.org/jira/browse/YARN-5781 Project: Hadoop YARN Issue Type: Improvement Reporter: Bibin A Chundatt As per the discussion in YARN-5773 with [~leftnoteasy]/[~sunilg] there is a scope of improvement for checking am limit based on partition and users. This jira is to track the same Comments 1: [wangda|https://issues.apache.org/jira/browse/YARN-5773?focusedCommentId=15602907&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15602907] Comment 2: [Sunil|https://issues.apache.org/jira/browse/YARN-5773?focusedCommentId=15607655&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15607655] -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5753) fix NPE in AMRMClientImpl.getMatchingRequests()
[ https://issues.apache.org/jira/browse/YARN-5753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607714#comment-15607714 ] Hudson commented on YARN-5753: -- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10683 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10683/]) YARN-5753. fix NPE in AMRMClientImpl.getMatchingRequests() (haibochen (rkanter: rev 44fdf009642ae4e99b15f89ec0ca43834f991ef3) * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/api/impl/AMRMClientImpl.java > fix NPE in AMRMClientImpl.getMatchingRequests() > --- > > Key: YARN-5753 > URL: https://issues.apache.org/jira/browse/YARN-5753 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn >Affects Versions: 3.0.0-alpha1 >Reporter: Haibo Chen >Assignee: Haibo Chen > Fix For: 3.0.0-alpha2 > > Attachments: yarn5753.001.patch, yarn5753.002.patch > > > {code:java} > RemoteRequestsTable remoteRequestsTable = getTable(0); > List> matchingRequests = > remoteRequestsTable.getMatchingRequests(priority, resourceName, > executionType, capability); > {code} > remoteRequestsTable can be null, causing NPE. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-4014) Support user cli interface in for Application Priority
[ https://issues.apache.org/jira/browse/YARN-4014?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607700#comment-15607700 ] stefanlee commented on YARN-4014: - it's done , i forgot to add |optional PriorityProto applicationPriority = 1;| in yarn_service_protos.proto > Support user cli interface in for Application Priority > -- > > Key: YARN-4014 > URL: https://issues.apache.org/jira/browse/YARN-4014 > Project: Hadoop YARN > Issue Type: Sub-task > Components: client, resourcemanager >Reporter: Rohith Sharma K S >Assignee: Rohith Sharma K S > Fix For: 2.8.0, 3.0.0-alpha1 > > Attachments: 0001-YARN-4014-V1.patch, 0001-YARN-4014.patch, > 0002-YARN-4014.patch, 0003-YARN-4014.patch, 0004-YARN-4014.patch, > 0004-YARN-4014.patch > > > Track the changes for user-RM client protocol i.e ApplicationClientProtocol > changes and discussions in this jira. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5575) Many classes use bare yarn. properties instead of the defined constants
[ https://issues.apache.org/jira/browse/YARN-5575?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607666#comment-15607666 ] Hudson commented on YARN-5575: -- SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #10682 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/10682/]) YARN-5575. Many classes use bare yarn. properties instead of the defined (aajisaka: rev d3bb69a66776e9f410150c4030ddb15981f58fb9) * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/src/test/java/org/apache/hadoop/yarn/applications/distributedshell/TestDistributedShell.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/ahs/TestRMApplicationHistoryWriter.java * (edit) hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/v2/TestMRJobs.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMProxyUsersConf.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestCapacitySchedulerQueueACLs.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairSchedulerQueueACLs.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/ReservationACLsTestBase.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesReservation.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRM.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestLeafQueue.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesHttpStaticUserPermissions.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMAdminService.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/test/java/org/apache/hadoop/yarn/server/timeline/security/TestTimelineAuthenticationFilterInitializer.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairSchedulerEventLog.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestApplicationLimits.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/src/test/java/org/apache/hadoop/yarn/applications/distributedshell/TestDistributedShellWithNodeLabels.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/monitor/capacity/TestProportionalCapacityPreemptionPolicy.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesAppsModification.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestReservations.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesDelegationTokenAuthentication.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestAppManager.java * (edit) hadoop-tools/hadoop-gridmix/src/test/java/org/apache/hadoop/mapred/gridmix/GridmixTestUtils.java * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server
[jira] [Commented] (YARN-5773) RM recovery too slow due to LeafQueue#activateApplication()
[ https://issues.apache.org/jira/browse/YARN-5773?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607655#comment-15607655 ] Sunil G commented on YARN-5773: --- In scheduler, as of now there are no event/apis to know whether recovery is done or not. Basically apps could be submitted even when none of the nodes are registered. I understood that the easy fix here is to skip invoking {{activateApplications}} during recovery. Its already have no meaning in the recovery flow. So as I mentioned earlier and as in patch, we can have this. Now the question here is to invoke {{activateApplications}} after scheduler recovery is done. Advantage of such a call is that scheduler will no longer need to worry about the chain of sequential steps of RM recovery (starting active service as last in those steps or not). Since recovery is done event based, a direct api will not be correct. Rather, one more event to be published. I will check the feasibility for this and update here. > RM recovery too slow due to LeafQueue#activateApplication() > --- > > Key: YARN-5773 > URL: https://issues.apache.org/jira/browse/YARN-5773 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Critical > Attachments: YARN-5773.0001.patch, YARN-5773.0002.patch, > YARN-5773.003.patch > > > # Submit application 10K application to default queue. > # All applications are in accepted state > # Now restart resourcemanager > For each application recovery {{LeafQueue#activateApplications()}} is > invoked.Resulting in AM limit check to be done even before Node managers are > getting registered. > Total iteration for N application is about {{N(N+1)/2}} for {{10K}} > application {{5000}} iterations causing time take for Rm to be active > more than 10 min. > Since NM resources are not yet added to during recovery we should skip > {{activateApplicaiton()}} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5753) fix NPE in AMRMClientImpl.getMatchingRequests()
[ https://issues.apache.org/jira/browse/YARN-5753?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15607657#comment-15607657 ] Robert Kanter commented on YARN-5753: - +1 > fix NPE in AMRMClientImpl.getMatchingRequests() > --- > > Key: YARN-5753 > URL: https://issues.apache.org/jira/browse/YARN-5753 > Project: Hadoop YARN > Issue Type: Bug > Components: yarn >Affects Versions: 3.0.0-alpha1 >Reporter: Haibo Chen >Assignee: Haibo Chen > Attachments: yarn5753.001.patch, yarn5753.002.patch > > > {code:java} > RemoteRequestsTable remoteRequestsTable = getTable(0); > List> matchingRequests = > remoteRequestsTable.getMatchingRequests(priority, resourceName, > executionType, capability); > {code} > remoteRequestsTable can be null, causing NPE. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org