Hi, Thanks for doing the update. Let's see if I contribute a few more times - if it becomes a pain for you / others to gatekeep me, we can revisit access then.
Cheers, On 08/06/16 13:49, Darin Johnson wrote: > Will do today, if you'd like to help with the documentation I could give > you access. > > On Wed, Jun 8, 2016 at 3:14 AM, Stephen Gran <stephen.g...@piksel.com> > wrote: > >> Hi, >> >> Can someone with access please correct the screenshot here: >> https://cwiki.apache.org/confluence/display/MYRIAD/Fine-grained+Scaling >> >> This gives the strong impression that you don't need an NM with non-zero >> resources. I think this is what initially steered me down the wrong path. >> >> Cheers, >> >> On 03/06/16 16:38, Darin Johnson wrote: >>> That is correct you need at least one node manager with the minimum >>> requirements to launch an ApplicationMaster. Otherwise YARN will throw >> an >>> exception. >>> >>> On Fri, Jun 3, 2016 at 10:52 AM, yuliya Feldman >> <yufeld...@yahoo.com.invalid >>>> wrote: >>> >>>> I believe you need at least one NM that is not subject to fine grain >>>> scaling. >>>> So far if total resources on the cluster is less then a single container >>>> needs for AM you won't be able to submit any app.As exception below >> tells >>>> you. >>>> (Invalid resource request, requested memory < 0, or requested memory >>> max >>>> configured, requestedMemory=1536, maxMemory=0 >>>> at) >>>> I believe by default when starting Myriad cluster one NM with non 0 >>>> capacity should start by default. >>>> In addition see in RM log whether offers with resources are coming to >> RM - >>>> this info should be in the log. >>>> >>>> From: Stephen Gran <stephen.g...@piksel.com> >>>> To: "dev@myriad.incubator.apache.org" < >> dev@myriad.incubator.apache.org> >>>> Sent: Friday, June 3, 2016 1:29 AM >>>> Subject: problem getting fine grained scaling workig >>>> >>>> Hi, >>>> >>>> I'm trying to get fine grained scaling going on a test mesos cluster. I >>>> have a single master and 2 agents. I am running 2 node managers with >>>> the zero profile, one per agent. I can see both of them in the RM UI >>>> reporting correctly as having 0 resources. >>>> >>>> I'm getting stack traces when I try to launch a sample application, >>>> though. I feel like I'm just missing something obvious somewhere - can >>>> anyone shed any light? >>>> >>>> This is on a build of yesterday's git head. >>>> >>>> Cheers, >>>> >>>> root@master:/srv/apps/hadoop# bin/yarn jar >>>> share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.2.jar teragen 10000 >>>> /outDir >>>> 16/06/03 08:23:33 INFO client.RMProxy: Connecting to ResourceManager at >>>> master.testing.local/10.0.5.3:8032 >>>> 16/06/03 08:23:34 INFO terasort.TeraSort: Generating 10000 using 2 >>>> 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: number of splits:2 >>>> 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: Submitting tokens for >>>> job: job_1464902078156_0001 >>>> 16/06/03 08:23:35 INFO mapreduce.JobSubmitter: Cleaning up the staging >>>> area /tmp/hadoop-yarn/staging/root/.staging/job_1464902078156_0001 >>>> java.io.IOException: >>>> org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: >>>> Invalid resource request, requested memory < 0, or requested memory > >>>> max configured, requestedMemory=1536, maxMemory=0 >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) >>>> at >>>> >>>> >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> >>>> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) >>>> >>>> at >>>> org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:306) >>>> at >>>> >>>> >> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:240) >>>> at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) >>>> at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> >>>> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) >>>> at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) >>>> at >> org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) >>>> at >>>> org.apache.hadoop.examples.terasort.TeraGen.run(TeraGen.java:301) >>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >>>> at >>>> org.apache.hadoop.examples.terasort.TeraGen.main(TeraGen.java:305) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> >>>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> at >>>> >>>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:497) >>>> at >>>> >>>> >> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:71) >>>> at >> org.apache.hadoop.util.ProgramDriver.run(ProgramDriver.java:144) >>>> at >>>> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:74) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> >>>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> at >>>> >>>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:497) >>>> at org.apache.hadoop.util.RunJar.run(RunJar.java:221) >>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:136) >>>> Caused by: >>>> org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: >>>> Invalid resource request, requested memory < 0, or requested memory > >>>> max configured, requestedMemory=1536, maxMemory=0 >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) >>>> at >>>> >>>> >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> >>>> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) >>>> >>>> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >>>> Method) >>>> at >>>> >>>> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >>>> at >>>> >>>> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>> at >> java.lang.reflect.Constructor.newInstance(Constructor.java:422) >>>> at >>>> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:101) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:239) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> >>>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> at >>>> >>>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:497) >>>> at >>>> >>>> >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191) >>>> at >>>> >>>> >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) >>>> at com.sun.proxy.$Proxy13.submitApplication(Unknown Source) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.submitApplication(YarnClientImpl.java:253) >>>> at >>>> >>>> >> org.apache.hadoop.mapred.ResourceMgrDelegate.submitApplication(ResourceMgrDelegate.java:290) >>>> at >>>> org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:290) >>>> ... 24 more >>>> Caused by: >>>> >>>> >> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException): >>>> Invalid resource request, requested memory < 0, or requested memory > >>>> max configured, requestedMemory=1536, maxMemory=0 >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) >>>> at >>>> >>>> >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> at >>>> >>>> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) >>>> >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1475) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:1412) >>>> at >>>> >>>> >> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) >>>> at com.sun.proxy.$Proxy12.submitApplication(Unknown Source) >>>> at >>>> >>>> >> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:236) >>>> ... 34 more >>>> >>>> >>>> Cheers, >>>> -- >>>> Stephen Gran >>>> Senior Technical Architect >>>> >>>> picture the possibilities | piksel.com >>>> This message is private and confidential. If you have received this >>>> message in error, please notify the sender or serviced...@piksel.com >> and >>>> remove it from your system. >>>> >>>> Piksel Inc is a company registered in the United States New York City, >>>> 1250 Broadway, Suite 1902, New York, NY 10001. F No. = 2931986 >>>> >>>> >>>> >>> >> >> -- >> Stephen Gran >> Senior Technical Architect >> >> picture the possibilities | piksel.com >> > -- Stephen Gran Senior Technical Architect picture the possibilities | piksel.com