Re: problem getting fine grained scaling working
That is normal behavior, Myriad keeps the resources to flexup a node manager incase a job comes in of a few seconds and then releases them. The info statement is arguably chatty and will probably go to debug in a few more releases. On Fri, Jun 3, 2016 at 9:18 AM, Stephen Gran wrote: > Hi, > > Not sure if this is relevant, but I see this in the RM logs: > > 2016-06-03 13:06:55,466 INFO > org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting > capacity for node slave1.testing.local to > 2016-06-03 13:06:55,467 INFO > > org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: > Update resource on node: slave1.testing.local from: vCores:0>, to: > 2016-06-03 13:06:55,467 INFO > org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting > capacity for node slave1.testing.local to > 2016-06-03 13:06:55,470 INFO > > org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: > Update resource on node: slave1.testing.local from: vCores:6>, to: > > > This is happening for each nodemanager, repeating every 5 or 6 seconds. > I'm assuming this will be the NM sending the actual capacity report to > the RM, for use in updating YARN's view of available resource. I don't > know if it should be going back and forth like it is, though? > > Cheers, > > On 03/06/16 09:29, Stephen Gran wrote: > > Hi, > > > > I'm trying to get fine grained scaling going on a test mesos cluster. I > > have a single master and 2 agents. I am running 2 node managers with > > the zero profile, one per agent. I can see both of them in the RM UI > > reporting correctly as having 0 resources. > > > > I'm getting stack traces when I try to launch a sample application, > > though. I feel like I'm just missing something obvious somewhere - can > > anyone shed any light? > > > > This is on a build of yesterday's git head. > > > > Cheers, > > > > root@master:/srv/apps/hadoop# bin/yarn jar > > share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.2.jar teragen 1 > > /outDir > > 16/06/03 08:23:33 INFO client.RMProxy: Connecting to ResourceManager at > > master.testing.local/10.0.5.3:8032 > > 16/06/03 08:23:34 INFO terasort.TeraSort: Generating 1 using 2 > > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: number of splits:2 > > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: Submitting tokens for > > job: job_1464902078156_0001 > > 16/06/03 08:23:35 INFO mapreduce.JobSubmitter: Cleaning up the staging > > area /tmp/hadoop-yarn/staging/root/.staging/job_1464902078156_0001 > > java.io.IOException: > > org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: > > Invalid resource request, requested memory < 0, or requested memory > > > max configured, requestedMemory=1536, maxMemory=0 > > at > > > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) > > at > > > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) > > at > > > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) > > at > > > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) > > at > > > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:422) > > at > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > > > > at > org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:306) > > at > > > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:240) > > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) > > a
Re: problem getting fine grained scaling working
Hi, Not sure if this is relevant, but I see this in the RM logs: 2016-06-03 13:06:55,466 INFO org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting capacity for node slave1.testing.local to 2016-06-03 13:06:55,467 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: Update resource on node: slave1.testing.local from: , to: 2016-06-03 13:06:55,467 INFO org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting capacity for node slave1.testing.local to 2016-06-03 13:06:55,470 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: Update resource on node: slave1.testing.local from: , to: This is happening for each nodemanager, repeating every 5 or 6 seconds. I'm assuming this will be the NM sending the actual capacity report to the RM, for use in updating YARN's view of available resource. I don't know if it should be going back and forth like it is, though? Cheers, On 03/06/16 09:29, Stephen Gran wrote: > Hi, > > I'm trying to get fine grained scaling going on a test mesos cluster. I > have a single master and 2 agents. I am running 2 node managers with > the zero profile, one per agent. I can see both of them in the RM UI > reporting correctly as having 0 resources. > > I'm getting stack traces when I try to launch a sample application, > though. I feel like I'm just missing something obvious somewhere - can > anyone shed any light? > > This is on a build of yesterday's git head. > > Cheers, > > root@master:/srv/apps/hadoop# bin/yarn jar > share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.2.jar teragen 1 > /outDir > 16/06/03 08:23:33 INFO client.RMProxy: Connecting to ResourceManager at > master.testing.local/10.0.5.3:8032 > 16/06/03 08:23:34 INFO terasort.TeraSort: Generating 1 using 2 > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: number of splits:2 > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: Submitting tokens for > job: job_1464902078156_0001 > 16/06/03 08:23:35 INFO mapreduce.JobSubmitter: Cleaning up the staging > area /tmp/hadoop-yarn/staging/root/.staging/job_1464902078156_0001 > java.io.IOException: > org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: > Invalid resource request, requested memory < 0, or requested memory > > max configured, requestedMemory=1536, maxMemory=0 > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > > at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:306) > at > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:240) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) > at org.apache.hadoop.examples.terasort.TeraGen.ru