[
https://issues.apache.org/jira/browse/SPARK-14389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15231070#comment-15231070
]
Steve Johnston commented on SPARK-14389:
----------------------------------------
Looking at stdout.txt (attached): in the stack trace it looks like
BroadcastNestedLoopJoin calls UnsafeRow.copy to initiate the failure.
>From the master node:
{code}jps -lvm |grep Submit{code}
{noformat}
30185 org.apache.spark.deploy.SparkSubmit
/home/hadoop/aqa_root/src/experimental/spark_test/sample_script.py -Xms1024M
-Xmx1024M -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
-XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
-XX:MaxHeapFreeRatio=70 -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
-XX:OnOutOfMemoryError=kill -9 %p
{noformat}
> OOM during BroadcastNestedLoopJoin
> ----------------------------------
>
> Key: SPARK-14389
> URL: https://issues.apache.org/jira/browse/SPARK-14389
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 1.6.0
> Environment: OS: Amazon Linux AMI 2015.09
> EMR: 4.3.0
> Hadoop: Amazon 2.7.1
> Spark 1.6.0
> Ganglia 3.7.2
> Master: m3.xlarge
> Core: m3.xlarge
> m3.xlarge: 4 CPU, 15GB mem, 2x40GB SSD
> Reporter: Steve Johnston
> Attachments: lineitem.tbl, plans.txt, sample_script.py, stdout.txt
>
>
> When executing attached sample_script.py in client mode with a single
> executor an exception occurs, "java.lang.OutOfMemoryError: Java heap space",
> during the self join of a small table, TPC-H lineitem generated for a 1M
> dataset. Also see execution log stdout.txt attached.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]