May be you can repartition?

2017-09-04 9:25 GMT+08:00 KhajaAsmath Mohammed <mdkhajaasm...@gmail.com>:

> Hi,
>
> I am getting java.lang.OutOfMemoryError: Java heap space error whenever I
> ran the spark sql job.
>
> I came to conclusion issue is because of reading number of files from
> spark.
>
> I am reading 37 partitions and each partition has around 2000 files with
> filesize more than 128 MB  37*2000 files from spark.
>
> can anyone provide solution on how to merge all files while reading in
> spark and run it efficently.
>
> Increasing executor memory didnt resolve my issue. I went from 16 GB to 64
> GB stil no luck.
>
> Thanks,
> Asmath
>

Reply via email to