[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15718750#comment-15718750 ] vinisha commented on MAPREDUCE-6514: can you please add affect version to the Jira. > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Blocker > Fix For: 2.7.3, 2.6.5, 3.0.0-alpha1 > > Attachments: MAPREDUCE-6514.01.patch, MAPREDUCE-6514.02.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15273665#comment-15273665 ] Varun Saxena commented on MAPREDUCE-6514: - Thanks [~leftnoteasy]. Sorry could not address the review comment from Vinod yesterday as it was late night here and waiting for QA report would have meant an even longer wait. > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Blocker > Fix For: 2.7.3, 2.6.5 > > Attachments: MAPREDUCE-6514.01.patch, MAPREDUCE-6514.02.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15273533#comment-15273533 ] Hudson commented on MAPREDUCE-6514: --- FAILURE: Integrated in Hadoop-trunk-Commit #9728 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/9728/]) MAPREDUCE-6514. Fixed MapReduce ApplicationMaster to properly updated (vinodkv: rev 8d48266720dcf0e71cfd87fef18b60a53aa1bef9) * hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/RMContainerRequestor.java * hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/test/java/org/apache/hadoop/mapreduce/v2/app/rm/TestRMContainerAllocator.java * hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/rm/RMContainerAllocator.java > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Blocker > Attachments: MAPREDUCE-6514.01.patch, MAPREDUCE-6514.02.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15273364#comment-15273364 ] Wangda Tan commented on MAPREDUCE-6514: --- [~vinodkv], This behavior is commented by [~rohithsharma] above in above comment, bq. As a result, number of containers count in the ask is increased as explained in the below... And because of MAPREDUCE-6302, it is possible that MR AM cancel all reducer requests and re-add all reducer requests at the same heartbeat, so the #containers increases fast in RM side and finally it becomes a enormous number. Since MAPREDUCE-6302 is included by branch-2.6/2.7, we need to back port this patch to branch-2.6/2.7 as well. > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Blocker > Attachments: MAPREDUCE-6514.01.patch, MAPREDUCE-6514.02.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15273338#comment-15273338 ] Vinod Kumar Vavilapalli commented on MAPREDUCE-6514: The patch looks good to me now. Tx [~leftnoteasy] for finishing the patch on behalf of [~varun_saxena]! BTW, why is this a blocker on the 2.6 / 2.7 maint lines? > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Blocker > Attachments: MAPREDUCE-6514.01.patch, MAPREDUCE-6514.02.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: mapreduce-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: mapreduce-issues-h...@hadoop.apache.org
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15272856#comment-15272856 ] Hadoop QA commented on MAPREDUCE-6514: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 11s {color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 7s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 20s {color} | {color:green} trunk passed with JDK v1.8.0_91 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 22s {color} | {color:green} trunk passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 17s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 28s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 46s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 15s {color} | {color:green} trunk passed with JDK v1.8.0_91 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 18s {color} | {color:green} trunk passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 23s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 17s {color} | {color:green} the patch passed with JDK v1.8.0_91 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 17s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 20s {color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 20s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 14s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 25s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 11s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} Patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 53s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 13s {color} | {color:green} the patch passed with JDK v1.8.0_91 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 15s {color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 13s {color} | {color:green} hadoop-mapreduce-client-app in the patch passed with JDK v1.8.0_91. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 49s {color} | {color:green} hadoop-mapreduce-client-app in the patch passed with JDK v1.7.0_95. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 21s {color} | {color:green} Patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 31m 47s {color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:cf2ee45 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12802482/MAPREDUCE-6514.02.patch | | JIRA Issue | MAPREDUCE-6514 | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 134bb84d3e93 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / d0da132 | | Default Java | 1.7.0_95
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14982766#comment-14982766 ] Varun Saxena commented on MAPREDUCE-6514: - Test failure to be handled by YARN-4320 > Job hangs as ask is not updated after ramping down of all reducers > -- > > Key: MAPREDUCE-6514 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-6514 > Project: Hadoop Map/Reduce > Issue Type: Bug > Components: applicationmaster >Affects Versions: 2.7.1 >Reporter: Varun Saxena >Assignee: Varun Saxena >Priority: Critical > Attachments: MAPREDUCE-6514.01.patch > > > In RMContainerAllocator#preemptReducesIfNeeded, we simply clear the scheduled > reduces map and put these reducers to pending. This is not updated in ask. So > RM keeps on assigning and AM is not able to assign as no reducer is > scheduled(check logs below the code). > If this is updated immediately, RM will be able to schedule mappers > immediately which anyways is the intention when we ramp down reducers. > Scheduler need not allocate for ramped down reducers > This if not handled can lead to map starvation as pointed out in > MAPREDUCE-6513 > {code} > LOG.info("Ramping down all scheduled reduces:" > + scheduledRequests.reduces.size()); > for (ContainerRequest req : scheduledRequests.reduces.values()) { > pendingReduces.add(req); > } > scheduledRequests.reduces.clear(); > {code} > {noformat} > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000215 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000216, > NodeId: hdszzdcxdat6g06u04p:26009, NodeHttpAddress: > hdszzdcxdat6g06u04p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.236:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Container not > assigned : container_1437451211867_1485_01_000216 > 2015-10-13 04:55:04,912 INFO [RMCommunicator Allocator] > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: Cannot assign > container Container: [ContainerId: container_1437451211867_1485_01_000217, > NodeId: hdszzdcxdat6g06u06p:26009, NodeHttpAddress: > hdszzdcxdat6g06u06p:26010, Resource: , Priority: 10, > Token: Token { kind: ContainerToken, service: 10.2.33.239:26009 }, ] for a > reduce as either container memory less than required 4096 or no pending > reduce tasks - reduces.isEmpty=true > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332)
[jira] [Commented] (MAPREDUCE-6514) Job hangs as ask is not updated after ramping down of all reducers
[ https://issues.apache.org/jira/browse/MAPREDUCE-6514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14981359#comment-14981359 ] Hadoop QA commented on MAPREDUCE-6514: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 6s {color} | {color:blue} docker + precommit patch detected. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s {color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s {color} | {color:green} The patch appears to include 1 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 3m 30s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 17s {color} | {color:green} trunk passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 19s {color} | {color:green} trunk passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 12s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 45s {color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 16s {color} | {color:green} trunk passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 17s {color} | {color:green} trunk passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 21s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 15s {color} | {color:green} the patch passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 15s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 17s {color} | {color:green} the patch passed with JDK v1.7.0_79 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 17s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 10s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 13s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s {color} | {color:green} Patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 50s {color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 14s {color} | {color:green} the patch passed with JDK v1.8.0_60 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 18s {color} | {color:green} the patch passed with JDK v1.7.0_79 {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 9m 42s {color} | {color:red} hadoop-mapreduce-client-app in the patch failed with JDK v1.8.0_60. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 10m 9s {color} | {color:red} hadoop-mapreduce-client-app in the patch failed with JDK v1.7.0_79. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 22s {color} | {color:red} Patch generated 7 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 29m 45s {color} | {color:black} {color} | \\ \\ || Reason || Tests || | JDK v1.7.0_79 Failed junit tests | hadoop.mapreduce.v2.app.TestRecovery | | | hadoop.mapreduce.jobhistory.TestJobHistoryEventHandler | | | hadoop.mapreduce.jobhistory.TestJobHistoryEventHandler | \\ \\ || Subsystem || Report/Notes || | Docker | Client=1.7.1 Server=1.7.1 Image:test-patch-base-hadoop-date2015-10-29 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12769614/MAPREDUCE-6514.01.patch | | JIRA Issue | MAPREDUCE-6514 | | Optional Tests | asflicense javac javadoc mvninstall unit findbugs checkstyle compile | | uname | Linux 1f4486852554 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /home/jenkins/jenkins-slave/workspace/PreCommit-MAPREDUCE-Build/patchprocess/apache-yetus-c3a2069/precommit/personality/hadoop.sh | | git revis