Hi Ram,

I see in the cluster yarn-site.xml, below two properties are configured with 
below settings..

yarn.scheduler.minimum-allocation-mb ===> 1024
yarn.scheduler.maximum-allocation-mb ===> 32768


So with the above settings at cluster level, I can’t increase the memory 
allocated for my DAG ?  Is there is any other way, I can increase the memory ?


Thanks a lot.


Regards,
Raja.

From: Munagala Ramanath <[email protected]<mailto:[email protected]>>
Reply-To: "[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Date: Tuesday, July 12, 2016 at 9:31 AM
To: "[email protected]<mailto:[email protected]>" 
<[email protected]<mailto:[email protected]>>
Subject: Re: DAG is failing due to memory issues

Please see: http://docs.datatorrent.com/troubleshooting/#configuring-memory

Ram

On Tue, Jul 12, 2016 at 6:57 AM, Raja.Aravapalli 
<[email protected]<mailto:[email protected]>> wrote:

Hi,

My DAG is failing with memory issues for container. Seeing below information in 
the log.



Diagnostics: Container [pid=xxx,containerID=container_xyclksdjf] is running 
beyond physical memory limits. Current usage: 1.0 GB of 1 GB physical memory 
used; 2.9 GB of 2.1 GB virtual memory used. Killing container.


Can someone help me on how I can fix this issue. Thanks a lot.



Regards,
Raja.

Reply via email to