Raja, Please increase the container size and launch the app again. yarn.scheduler .maximum-allocation-mb is for the container and not for the DAG and the error message showed by you is for the container.
Here is one quick way, use the following attribute. <property> <name>dt.operator.*.attr.MEMORY_MB</name> <value>4096</value> </property> On Tue, Jul 12, 2016 at 9:24 AM Raja.Aravapalli <[email protected]> wrote: > > Hi Ram, > > Sorry I did not share that details of 32gb with you. > > I am saying 32gb is allocated because, I observed the same on UI, when the > application is running. But now, as the DAG is failed, I cannot take a > screenshot and send!! > > > Regards, > Raja. > > From: Munagala Ramanath <[email protected]> > Reply-To: "[email protected]" <[email protected]> > Date: Tuesday, July 12, 2016 at 11:06 AM > > To: "[email protected]" <[email protected]> > Subject: Re: DAG is failing due to memory issues > > How do you know it is allocating 32GB ? The diagnostic message you posted > does not show > that. > > Ram > > On Tue, Jul 12, 2016 at 8:51 AM, Raja.Aravapalli < > [email protected]> wrote: > >> >> Thanks for the response Sandesh. >> >> Since our yarn-site is configured with value *32768* for the property * >> yarn.scheduler.maximum-allocation-mb*, it is allocating a max of *32gb* >> and not more than that!! >> >> >> Wish to know, is there a way I can increase the max allowed value ? OR, >> since it is configured in yarn-site.xml, I *cannot* increase it ? >> >> >> >> Regards, >> Raja. >> >> From: Sandesh Hegde <[email protected]> >> Reply-To: "[email protected]" <[email protected]> >> Date: Tuesday, July 12, 2016 at 10:46 AM >> To: "[email protected]" <[email protected]> >> Subject: Re: DAG is failing due to memory issues >> >> Quoting from the doc shared by the Ram, those parameters control operator >> memory size. >> >> actual container memory allocated by RM has to lie between >> >> [yarn.scheduler.minimum-allocation-mb, yarn.scheduler.maximum-allocation-mb] >> >> >> On Tue, Jul 12, 2016 at 8:38 AM Raja.Aravapalli < >> [email protected]> wrote: >> >>> >>> Hi Ram, >>> >>> I see in the cluster yarn-site.xml, below two properties are configured >>> with below settings.. >>> >>> yarn.scheduler.minimum-allocation-mb ===> 1024 >>> yarn.scheduler.maximum-allocation-mb ===> 32768 >>> >>> >>> So with the above settings at cluster level, I can’t increase the memory >>> allocated for my DAG ? Is there is any other way, I can increase the >>> memory ? >>> >>> >>> Thanks a lot. >>> >>> >>> Regards, >>> Raja. >>> >>> From: Munagala Ramanath <[email protected]> >>> Reply-To: "[email protected]" <[email protected]> >>> Date: Tuesday, July 12, 2016 at 9:31 AM >>> To: "[email protected]" <[email protected]> >>> Subject: Re: DAG is failing due to memory issues >>> >>> Please see: >>> http://docs.datatorrent.com/troubleshooting/#configuring-memory >>> >>> Ram >>> >>> On Tue, Jul 12, 2016 at 6:57 AM, Raja.Aravapalli < >>> [email protected]> wrote: >>> >>>> >>>> Hi, >>>> >>>> My DAG is failing with memory issues for container. Seeing below >>>> information in the log. >>>> >>>> >>>> >>>> Diagnostics: Container [pid=xxx,containerID=container_xyclksdjf] is >>>> running beyond physical memory limits. Current usage: 1.0 GB of 1 GB >>>> physical memory used; 2.9 GB of 2.1 GB virtual memory used. Killing >>>> container. >>>> >>>> >>>> Can someone help me on how I can fix this issue. Thanks a lot. >>>> >>>> >>>> >>>> Regards, >>>> Raja. >>>> >>> >>> >
