Hi, OK. That helps, thank you. I think I just misunderstood the docs (or they never said explicitly that you did need at least some static resource), and I scaled down the initial nm.medium that got started. I get a bit further now, and jobs start but are killed with:
Diagnostics: Container [pid=3865,containerID=container_1465112239753_0001_03_000001] is running beyond virtual memory limits. Current usage: 50.7 MB of 0B physical memory used; 2.6 GB of 0B virtual memory used. Killing container When I've seen this in the past with yarn but without myriad, it was usually about ratios of vmem to mem and things like that - I've tried some of those knobs, but I didn't expect much result and didn't get any. What strikes me about the error message is that the vmem and mem allocations are for 0. I'm sorry for asking what are probably naive questions here, I couldn't find a different forum. If there is one, please point me there so I don't disrupt the dev flow here. I can see this in the logs: 2016-06-05 07:39:25,687 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_1465112239753_0001_03_000001 Container Transitioned from NEW to ALLOCATED 2016-06-05 07:39:25,688 INFO org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=root OPERATION=AM Allocated Container TARGET=SchedulerApp RESULT=SUCCESS APPID=application_1465112239753_0001 CONTAINERID=container_1465112239753_0001_03_000001 2016-06-05 07:39:25,688 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode: Assigned container container_1465112239753_0001_03_000001 of capacity <memory:0, vCores:0> on host slave2.testing.local:26688, which has 1 containers, <memory:0, vCores:0> used and <memory:4096, vCores:1> available after allocation 2016-06-05 07:39:25,689 INFO org.apache.hadoop.yarn.server.resourcemanager.security.NMTokenSecretManagerInRM: Sending NMToken for nodeId : slave2.testing.local:26688 for container : container_1465112239753_0001_03_000001 2016-06-05 07:39:25,696 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_1465112239753_0001_03_000001 Container Transitioned from ALLOCATED to ACQUIRED 2016-06-05 07:39:25,696 INFO org.apache.hadoop.yarn.server.resourcemanager.security.NMTokenSecretManagerInRM: Clear node set for appattempt_1465112239753_0001_000003 2016-06-05 07:39:25,696 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: Storing attempt: AppId: application_1465112239753_0001 AttemptId: appattempt_1465112239753_0001_000003 MasterContainer: Container: [ContainerId: container_1465112239753_0001_03_000001, NodeId: slave2.testing.local:26688, NodeHttpAddress: slave2.testing.local:24387, Resource: <memory:0, vCores:0>, Priority: 0, Token: Token { kind: ContainerToken, service: 10.0.5.5:26688 }, ] 2016-06-05 07:39:25,697 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: appattempt_1465112239753_0001_000003 State change from SCHEDULED to ALLOCATED_SAVING 2016-06-05 07:39:25,698 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: appattempt_1465112239753_0001_000003 State change from ALLOCATED_SAVING to ALLOCATED 2016-06-05 07:39:25,699 INFO org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: Launching masterappattempt_1465112239753_0001_000003 2016-06-05 07:39:25,705 INFO org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: Setting up container Container: [ContainerId: container_1465112239753_0001_03_000001, NodeId: slave2.testing.local:26688, NodeHttpAddress: slave2.testing.local:24387, Resource: <memory:0, vCores:0>, Priority: 0, Token: Token { kind: ContainerToken, service: 10.0.5.5:26688 }, ] for AM appattempt_1465112239753_0001_000003 2016-06-05 07:39:25,705 INFO org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: Command to launch container container_1465112239753_0001_03_000001 : $JAVA_HOME/bin/java -Djava.io.tmpdir=$PWD/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=<LOG_DIR> -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA -Dhadoop.root.logfile=syslog -Xmx1024m org.apache.hadoop.mapreduce.v2.app.MRAppMaster 1><LOG_DIR>/stdout 2><LOG_DIR>/stderr 2016-06-05 07:39:25,706 INFO org.apache.hadoop.yarn.server.resourcemanager.security.AMRMTokenSecretManager: Create AMRMToken for ApplicationAttempt: appattempt_1465112239753_0001_000003 2016-06-05 07:39:25,707 INFO org.apache.hadoop.yarn.server.resourcemanager.security.AMRMTokenSecretManager: Creating password for appattempt_1465112239753_0001_000003 2016-06-05 07:39:25,727 INFO org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher: Done launching container Container: [ContainerId: container_1465112239753_0001_03_000001, NodeId: slave2.testing.local:26688, NodeHttpAddress: slave2.testing.local:24387, Resource: <memory:0, vCores:0>, Priority: 0, Token: Token { kind: ContainerToken, service: 10.0.5.5:26688 }, ] for AM appattempt_1465112239753_0001_000003 2016-06-05 07:39:25,728 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: appattempt_1465112239753_0001_000003 State change from ALLOCATED to LAUNCHED 2016-06-05 07:39:25,736 WARN org.apache.myriad.scheduler.event.handlers.StatusUpdateEventHandler: Task: yarn_container_1465112239753_0001_03_000001 not found, status: TASK_RUNNING 2016-06-05 07:39:26,510 INFO org.apache.hadoop.yarn.util.RackResolver: Resolved slave1.testing.local to /default-rack 2016-06-05 07:39:26,517 WARN org.apache.myriad.scheduler.fgs.NMHeartBeatHandler: FineGrainedScaling feature got invoked for a NM with non-zero capacity. Host: slave1.testing.local, Mem: 4096, CPU: 0. Setting the NM's capacity to (0G,0CPU) 2016-06-05 07:39:26,517 INFO org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeImpl: slave1.testing.local:29121 Node Transitioned from NEW to RUNNING 2016-06-05 07:39:26,518 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler: Added node slave1.testing.local:29121 cluster capacity: <memory:4096, vCores:1> 2016-06-05 07:39:26,519 INFO org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: afterSchedulerEventHandled: NM registration from node slave1.testing.local 2016-06-05 07:39:26,528 INFO org.apache.hadoop.yarn.server.resourcemanager.ResourceTrackerService: received container statuses on node manager register :[container_id { app_attempt_id { application_id { id: 1 cluster_timestamp: 1465112239753 } attemptId: 2 } id: 1 } container_state: C_RUNNING resource { memory: 0 virtual_cores: 0 } priority { priority: 0 } diagnostics: "" container_exit_status: -1000 creation_time: 1465112356478] 2016-06-05 07:39:26,530 INFO org.apache.hadoop.yarn.server.resourcemanager.ResourceTrackerService: NodeManager from node slave1.testing.local(cmPort: 29121 httpPort: 20456) registered with capability: <memory:0, vCores:0>, assigned nodeId slave1.testing.local:29121 2016-06-05 07:39:26,611 INFO org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting capacity for node slave1.testing.local to <memory:4637, vCores:6> 2016-06-05 07:39:26,611 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: Update resource on node: slave1.testing.local from: <memory:0, vCores:0>, to: <memory:4637, vCores:6> 2016-06-05 07:39:26,615 INFO org.apache.myriad.scheduler.fgs.YarnNodeCapacityManager: Setting capacity for node slave1.testing.local to <memory:0, vCores:0> 2016-06-05 07:39:26,616 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.AbstractYarnScheduler: Update resource on node: slave1.testing.local from: <memory:4637, vCores:6>, to: <memory:0, vCores:0> 2016-06-05 07:39:26,691 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_1465112239753_0001_03_000001 Container Transitioned from ACQUIRED to RUNNING 2016-06-05 07:39:26,835 WARN org.apache.myriad.scheduler.event.handlers.StatusUpdateEventHandler: Task: yarn_container_1465112239753_0001_03_000001 not found, status: TASK_FINISHED 2016-06-05 07:39:27,603 INFO org.apache.myriad.scheduler.event.handlers.ResourceOffersEventHandler: Received offers 1 2016-06-05 07:39:27,748 INFO org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: container_1465112239753_0001_03_000001 Container Transitioned from RUNNING to COMPLETED 2016-06-05 07:39:27,748 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FSAppAttempt: Completed container: container_1465112239753_0001_03_000001 in state: COMPLETED event:FINISHED 2016-06-05 07:39:27,748 INFO org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=root OPERATION=AM Released Container TARGET=SchedulerApp RESULT=SUCCESS APPID=application_1465112239753_0001 CONTAINERID=container_1465112239753_0001_03_000001 2016-06-05 07:39:27,748 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode: Released container container_1465112239753_0001_03_000001 of capacity <memory:0, vCores:0> on host slave2.testing.local:26688, which currently has 0 containers, <memory:0, vCores:0> used and <memory:4096, vCores:1> available, release resources=true 2016-06-05 07:39:27,748 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler: Application attempt appattempt_1465112239753_0001_000003 released container container_1465112239753_0001_03_000001 on node: host: slave2.testing.local:26688 #containers=0 available=<memory:4096, vCores:1> used=<memory:0, vCores:0> with event: FINISHED 2016-06-05 07:39:27,749 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: Updating application attempt appattempt_1465112239753_0001_000003 with final state: FAILED, and exit status: -103 2016-06-05 07:39:27,750 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: appattempt_1465112239753_0001_000003 State change from LAUNCHED to FINAL_SAVING 2016-06-05 07:39:27,751 INFO org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService: Unregistering app attempt : appattempt_1465112239753_0001_000003 2016-06-05 07:39:27,751 INFO org.apache.hadoop.yarn.server.resourcemanager.security.AMRMTokenSecretManager: Application finished, removing password for appattempt_1465112239753_0001_000003 2016-06-05 07:39:27,751 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptImpl: appattempt_1465112239753_0001_000003 State change from FINAL_SAVING to FAILED 2016-06-05 07:39:27,751 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: The number of failed attempts is 2. The max attempts is 2 2016-06-05 07:39:27,753 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: Updating application application_1465112239753_0001 with final state: FAILED 2016-06-05 07:39:27,756 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: application_1465112239753_0001 State change from ACCEPTED to FINAL_SAVING 2016-06-05 07:39:27,757 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler: Application appattempt_1465112239753_0001_000003 is done. finalState=FAILED 2016-06-05 07:39:27,757 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo: Application application_1465112239753_0001 requests cleared 2016-06-05 07:39:27,758 INFO org.apache.hadoop.yarn.server.resourcemanager.recovery.RMStateStore: Updating info for app: application_1465112239753_0001 2016-06-05 07:39:27,758 INFO org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppImpl: Application application_1465112239753_0001 failed 2 times due to AM Container for appattempt_1465112239753_0001_000003 exited with exitCode: -103 For more detailed output, check application tracking page:http://master.testing.local:8088/cluster/app/application_1465112239753_0001Then, click on links to logs of each attempt. Diagnostics: Container [pid=3865,containerID=container_1465112239753_0001_03_000001] is running beyond virtual memory limits. Current usage: 50.7 MB of 0B physical memory used; 2.6 GB of 0B virtual memory used. Killing container. Dump of the process-tree for container_1465112239753_0001_03_000001 : |- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE |- 3873 3865 3865 3865 (java) 80 26 2770927616 12614 /usr/lib/jvm/java-8-openjdk-amd64/bin/java -Djava.io.tmpdir=/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/application_1465112239753_0001/container_1465112239753_0001_03_000001/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/srv/apps/hadoop-2.7.2/logs/userlogs/application_1465112239753_0001/container_1465112239753_0001_03_000001 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA -Dhadoop.root.logfile=syslog -Xmx1024m org.apache.hadoop.mapreduce.v2.app.MRAppMaster |- 3865 3863 3865 3865 (bash) 0 1 11427840 354 /bin/bash -c /usr/lib/jvm/java-8-openjdk-amd64/bin/java -Djava.io.tmpdir=/tmp/hadoop-root/nm-local-dir/usercache/root/appcache/application_1465112239753_0001/container_1465112239753_0001_03_000001/tmp -Dlog4j.configuration=container-log4j.properties -Dyarn.app.container.log.dir=/srv/apps/hadoop-2.7.2/logs/userlogs/application_1465112239753_0001/container_1465112239753_0001_03_000001 -Dyarn.app.container.log.filesize=0 -Dhadoop.root.logger=INFO,CLA -Dhadoop.root.logfile=syslog -Xmx1024m org.apache.hadoop.mapreduce.v2.app.MRAppMaster 1>/srv/apps/hadoop-2.7.2/logs/userlogs/application_1465112239753_0001/container_1465112239753_0001_03_000001/stdout 2>/srv/apps/hadoop-2.7.2/logs/userlogs/application_1465112239753_0001/container_1465112239753_0001_03_000001/stderr Container killed on request. Exit code is 143 Container exited with a non-zero exit code 143 Failing this attempt. Failing the application. On 03/06/16 15:52, yuliya Feldman wrote: > I believe you need at least one NM that is not subject to fine grain scaling. > So far if total resources on the cluster is less then a single container > needs for AM you won't be able to submit any app.As exception below tells you. > (Invalid resource request, requested memory < 0, or requested memory >max > configured, requestedMemory=1536, maxMemory=0 > at) > I believe by default when starting Myriad cluster one NM with non 0 capacity > should start by default. > In addition see in RM log whether offers with resources are coming to RM - > this info should be in the log. > > From: Stephen Gran <stephen.g...@piksel.com> > To: "dev@myriad.incubator.apache.org" <dev@myriad.incubator.apache.org> > Sent: Friday, June 3, 2016 1:29 AM > Subject: problem getting fine grained scaling workig > > Hi, > > I'm trying to get fine grained scaling going on a test mesos cluster. I > have a single master and 2 agents. I am running 2 node managers with > the zero profile, one per agent. I can see both of them in the RM UI > reporting correctly as having 0 resources. > > I'm getting stack traces when I try to launch a sample application, > though. I feel like I'm just missing something obvious somewhere - can > anyone shed any light? > > This is on a build of yesterday's git head. > > Cheers, > > root@master:/srv/apps/hadoop# bin/yarn jar > share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.2.jar teragen 10000 > /outDir > 16/06/03 08:23:33 INFO client.RMProxy: Connecting to ResourceManager at > master.testing.local/10.0.5.3:8032 > 16/06/03 08:23:34 INFO terasort.TeraSort: Generating 10000 using 2 > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: number of splits:2 > 16/06/03 08:23:34 INFO mapreduce.JobSubmitter: Submitting tokens for > job: job_1464902078156_0001 > 16/06/03 08:23:35 INFO mapreduce.JobSubmitter: Cleaning up the staging > area /tmp/hadoop-yarn/staging/root/.staging/job_1464902078156_0001 > java.io.IOException: > org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: > Invalid resource request, requested memory < 0, or requested memory > > max configured, requestedMemory=1536, maxMemory=0 > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > > at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:306) > at > org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:240) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) > at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) > at org.apache.hadoop.examples.terasort.TeraGen.run(TeraGen.java:301) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > at org.apache.hadoop.examples.terasort.TeraGen.main(TeraGen.java:305) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:71) > at org.apache.hadoop.util.ProgramDriver.run(ProgramDriver.java:144) > at > org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:74) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at org.apache.hadoop.util.RunJar.run(RunJar.java:221) > at org.apache.hadoop.util.RunJar.main(RunJar.java:136) > Caused by: > org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException: > Invalid resource request, requested memory < 0, or requested memory > > max configured, requestedMemory=1536, maxMemory=0 > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:422) > at > org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) > at > org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:101) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:239) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) > at com.sun.proxy.$Proxy13.submitApplication(Unknown Source) > at > org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.submitApplication(YarnClientImpl.java:253) > at > org.apache.hadoop.mapred.ResourceMgrDelegate.submitApplication(ResourceMgrDelegate.java:290) > at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:290) > ... 24 more > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.InvalidResourceRequestException): > Invalid resource request, requested memory < 0, or requested memory > > max configured, requestedMemory=1536, maxMemory=0 > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.validateResourceRequest(SchedulerUtils.java:268) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:228) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerUtils.normalizeAndValidateRequest(SchedulerUtils.java:236) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.validateAndCreateResourceRequest(RMAppManager.java:385) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:329) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:281) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:580) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:218) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:419) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2045) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2043) > > at org.apache.hadoop.ipc.Client.call(Client.java:1475) > at org.apache.hadoop.ipc.Client.call(Client.java:1412) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) > at com.sun.proxy.$Proxy12.submitApplication(Unknown Source) > at > org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.submitApplication(ApplicationClientProtocolPBClientImpl.java:236) > ... 34 more > > > Cheers, > -- > Stephen Gran > Senior Technical Architect > > picture the possibilities | piksel.com > This message is private and confidential. If you have received this message > in error, please notify the sender or serviced...@piksel.com and remove it > from your system. > > Piksel Inc is a company registered in the United States New York City, 1250 > Broadway, Suite 1902, New York, NY 10001. F No. = 2931986 > > > -- Stephen Gran Senior Technical Architect picture the possibilities | piksel.com