Hi ,

Yes I understand its skew based problem but how can it be avoided . Could
you please suggest?

I am in Spark2.4

Thanks
Rajat

On Tue, Jan 26, 2021 at 3:58 PM German Schiavon <gschiavonsp...@gmail.com>
wrote:

> Hi,
>
> One word : SKEW
>
> It seems the classic skew problem, you would have to apply skew techniques
> to repartition your data properly or if you are in spark 3.0+ try the
> skewJoin optimization.
>
> On Tue, 26 Jan 2021 at 11:20, rajat kumar <kumar.rajat20...@gmail.com>
> wrote:
>
>> Hi Everyone,
>>
>> I am running a spark application where I have applied 2 left joins. 1st
>> join in Broadcast and another one is normal.
>> Out of 200 tasks , last 1 task is stuck . It is running at "ANY" Locality
>> level. It seems data skewness issue.
>> It is doing too much spill and shuffle write is too much. Following error
>> is coming in executor logs:
>>
>> INFO UnsafeExternalSorter: Thread spilling sort data of 10.4 GB to disk
>> (10  times so far)
>>
>>
>> Can anyone please suggest what can be wrong?
>>
>> Thanks
>> Rajat
>>
>

Reply via email to