[ 
https://issues.apache.org/jira/browse/SPARK-14389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15231070#comment-15231070
 ] 

Steve Johnston commented on SPARK-14389:
----------------------------------------

Looking at stdout.txt (attached): in the stack trace it looks like 
BroadcastNestedLoopJoin calls UnsafeRow.copy to initiate the failure.

>From the master node:
{code}jps -lvm |grep Submit{code}
{noformat}
30185 org.apache.spark.deploy.SparkSubmit 
/home/hadoop/aqa_root/src/experimental/spark_test/sample_script.py -Xms1024M 
-Xmx1024M -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties 
-XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 
-XX:MaxHeapFreeRatio=70 -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M 
-XX:OnOutOfMemoryError=kill -9 %p
{noformat}


> OOM during BroadcastNestedLoopJoin
> ----------------------------------
>
>                 Key: SPARK-14389
>                 URL: https://issues.apache.org/jira/browse/SPARK-14389
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 1.6.0
>         Environment: OS: Amazon Linux AMI 2015.09
> EMR: 4.3.0
> Hadoop: Amazon 2.7.1
> Spark 1.6.0
> Ganglia 3.7.2
> Master: m3.xlarge
> Core: m3.xlarge
> m3.xlarge: 4 CPU, 15GB mem, 2x40GB SSD
>            Reporter: Steve Johnston
>         Attachments: lineitem.tbl, plans.txt, sample_script.py, stdout.txt
>
>
> When executing attached sample_script.py in client mode with a single 
> executor an exception occurs, "java.lang.OutOfMemoryError: Java heap space", 
> during the self join of a small table, TPC-H lineitem generated for a 1M 
> dataset. Also see execution log stdout.txt attached.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to