The setting to increase is spark.yarn.executor.memoryOverhead

On Wed, Apr 15, 2015 at 6:35 AM, Brahma Reddy Battula <
brahmareddy.batt...@huawei.com> wrote:

> Hello Sean Owen,
>
> Thanks for your reply..I"ll increase overhead memory and check it..
>
>
> Bytheway ,Any difference between 1.1 and 1.2 makes, this issue..? Since It
> was passing spark 1.1 and throwing following error in 1.2...( this makes me
> doubt full)
>
>
>
> Thanks & Regards
> Brahma Reddy Battula
>
>
>
>
> ________________________________________
> From: Sean Owen [so...@cloudera.com]
> Sent: Wednesday, April 15, 2015 2:49 PM
> To: Brahma Reddy Battula
> Cc: Akhil Das; user@spark.apache.org
> Subject: Re: Running beyond physical memory limits
>
> This is not related to executor memory, but the extra overhead
> subtracted from the executor's size in order to avoid using more than
> the physical memory that YARN allows. That is, if you declare a 32G
> executor YARN lets you use 32G physical memory but your JVM heap must
> be significantly less than 32G max. This is the overhead factor that
> is subtracted for you, and it seems to need to be bigger in your case.
>
> On Wed, Apr 15, 2015 at 10:16 AM, Brahma Reddy Battula
> <brahmareddy.batt...@huawei.com> wrote:
> > Thanks lot for your reply..
> >
> >   There is no issue with spark1.1..Following issue came when I upgrade to
> > spark2.0...Hence I did not decrease spark.executor.memory...
> > I mean to say, used same config for spark1.1 and spark1.2..
> >
> > Is there any issue with spark1.2..?
> > Or Yarn will lead this..?
> > And why executor will not release memory, if there are tasks running..?
> >
> >
> > Thanks & Regards
> >
> > Brahma Reddy Battula
> >
> >
> > ________________________________
> > From: Akhil Das [ak...@sigmoidanalytics.com]
> > Sent: Wednesday, April 15, 2015 2:35 PM
> > To: Brahma Reddy Battula
> > Cc: user@spark.apache.org
> > Subject: Re: Running beyond physical memory limits
> >
> > Did you try reducing your spark.executor.memory?
> >
> > Thanks
> > Best Regards
> >
> > On Wed, Apr 15, 2015 at 2:29 PM, Brahma Reddy Battula
> > <brahmareddy.batt...@huawei.com> wrote:
> >>
> >> Hello Sparkers
> >>
> >>
> >> I am newbie to spark and  need help.. We are using spark 1.2, we are
> >> getting the following error and executor is getting killed..I seen
> >> SPARK-1930 and it should be in 1.2..
> >>
> >> Any pointer to following error, like what might lead this error..
> >>
> >>
> >> 2015-04-15 11:55:39,697 | WARN  | Container Monitor | Container
> >> [pid=126843,containerID=container_1429065217137_0012_01_-411041790] is
> >> running beyond physical memory limits. Current usage: 26.0 GB of 26 GB
> >> physical memory used; 26.7 GB of 260 GB virtual memory used. Killing
> >> container.
> >> Dump of the process-tree for container_1429065217137_0012_01_-411041790
> :
> >> |- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS)
> >> SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE
> >>         |- 126872 126843 126843 126843 (java) 2049457 22816 28673892352
> >> 6824864 /opt/huawei/Bigdata/jdk1.7.0_76//bin/java -server
> >> -XX:OnOutOfMemoryError=kill %p -Xms24576m -Xmx24576m
> >>
> -Dlog4j.configuration=file:/opt/huawei/Bigdata/DataSight_FM_BasePlatform_V100R001C00_Spark/spark/conf/log4j-executor.properties
> >>
> -Djava.library.path=/opt/huawei/Bigdata/DataSight_FM_BasePlatform_V100R001C00_Hadoop//hadoop/lib/native
> >>
> -Djava.io.tmpdir=/srv/BigData/hadoop/data4/nm/localdir/usercache/ossuser/appcache/application_1429065217137_0012/container_1429065217137_0012_01_-411041790/tmp
> >> -Dspark.driver.port=23204 -Dspark.random.port.max=23999
> >> -Dspark.akka.threads=32 -Dspark.akka.frameSize=10
> -Dspark.akka.timeout=100
> >> -Dspark.ui.port=23000 -Dspark.random.port.min=23000
> >>
> -Dspark.yarn.app.container.log.dir=/srv/BigData/hadoop/data5/nm/containerlogs/application_1429065217137_0012/container_1429065217137_0012_01_-411041790
> >> org.apache.spark.executor.CoarseGrainedExecutorBackend
> >> akka.tcp://sparkDriver@172.57.1.61:23204/user/CoarseGrainedScheduler 3
> >> hadoopc1h11 10 application_1429065217137_0012         |- 126843 76960
> 126843
> >> 126843 (bash) 0 0 11603968 331 /bin/bash -c
> >> /opt/huawei/Bigdata/jdk1.7.0_76//bin/java -server
> >> -XX:OnOutOfMemoryError='kill %p' -Xms24576m -Xmx24576m
> >>
> -Dlog4j.configuration=file:/opt/huawei/Bigdata/DataSight_FM_BasePlatform_V100R001C00_Spark/spark/conf/log4j-executor.properties
> >>
> -Djava.library.path=/opt/huawei/Bigdata/DataSight_FM_BasePlatform_V100R001C00_Hadoop//hadoop/lib/native
> >>
> -Djava.io.tmpdir=/srv/BigData/hadoop/data4/nm/localdir/usercache/ossuser/appcache/application_1429065217137_0012/container_1429065217137_0012_01_-411041790/tmp
> >> '-Dspark.driver.port=23204' '-Dspark.random.port.max=23999'
> >> '-Dspark.akka.threads=32' '-Dspark.akka.frameSize=10'
> >> '-Dspark.akka.timeout=100' '-Dspark.ui.port=23000'
> >> '-Dspark.random.port.min=23000'
> >>
> -Dspark.yarn.app.container.log.dir=/srv/BigData/hadoop/data5/nm/containerlogs/application_1429065217137_0012/container_1429065217137_0012_01_-411041790
> >> org.apache.spark.executor.CoarseGrainedExecutorBackend
> >> akka.tcp://sparkDriver@172.57.1.61:23204/user/CoarseGrainedScheduler 3
> >> hadoopc1h11 10 application_1429065217137_0012 1>
> >>
> /srv/BigData/hadoop/data5/nm/containerlogs/application_1429065217137_0012/container_1429065217137_0012_01_-411041790/stdout
> >> 2>
> >>
> /srv/BigData/hadoop/data5/nm/containerlogs/application_1429065217137_0012/container_1429065217137_0012_01_-411041790/stderr
> >>  |
> >>
> org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl$MonitoringThread.run(ContainersMonitorImpl.java:447)
> >>
> >>
> >>
> >> And some doubts
> >>
> >>
> >> 1) why executor will not release memory, if there are tasks running..?
> >>
> >>
> >>
> >> 2) is there issue from hadoop which will lead this error..?
> >>
> >>
> >>
> >> Any help , will be appreciated...
> >>
> >>
> >>
> >>
> >> Thanks & Regards
> >>
> >> Brahma Reddy Battula
> >>
> >>
> >>
> >>
> >
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to