Just a thought; this has worked for me before on standalone client
with a similar OOM error in a driver thread. Try setting:
export SPARK_DAEMON_MEMORY=4G #or whatever size you can afford on your machine
in your environment/spark-env.sh before running spark-submit.
Mike

On 9/2/15, ankit tyagi <ankittyagi.mn...@gmail.com> wrote:
> Hi All,
>
> I am using spark-sql 1.3.1 with hadoop 2.4.0 version.  I am running sql
> query against parquet files and wanted to save result on s3 but looks like
> https://issues.apache.org/jira/browse/SPARK-2984 problem still coming while
> saving data to s3.
>
> Hence Now i am saving result on hdfs and with the help
> of JavaSparkListener, copying file from hdfs to s3 with hadoop fileUtil
> in onApplicationEnd method. But  my job is getting failed with OOM in spark
> driver.
>
> *5/09/02 04:17:57 INFO cluster.YarnClusterSchedulerBackend: Asking each
> executor to shut down*
> *15/09/02 04:17:59 INFO
> scheduler.OutputCommitCoordinator$OutputCommitCoordinatorActor:
> OutputCommitCoordinator stopped!*
> *Exception in thread "Reporter" *
> *Exception: java.lang.OutOfMemoryError thrown from the
> UncaughtExceptionHandler in thread "Reporter"*
> *Exception in thread "SparkListenerBus" *
> *Exception: java.lang.OutOfMemoryError thrown from the
> UncaughtExceptionHandler in thread "SparkListenerBus"*
> *Exception in thread "Driver" *
> *Exception: java.lang.OutOfMemoryError thrown from the
> UncaughtExceptionHandler in thread "Driver"*
>
>
> Strage part is, result is getting saved on HDFS but while copying file job
> is getting failed. size of file is under 1MB.
>
> Any help or leads would be appreciated.
>


-- 
Thanks,
Mike

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to