Hi,
I see a similar behaviour in an exactly similar scenario at my deployment as
well. I am using scala, so the behaviour is not limited to pyspark.
In my observation 9 out of 10 partitions (as in my case) are of similar size
~38 GB each and final one is significantly larger ~59 GB.
Prime number
Hi,
I am using hiveContext.sql() method to select data from source table and
insert into parquet tables.
The query executed from spark takes about 3x more disk space to write
the same number of rows compared to when fired from impala.
Just wondering if this is normal behaviour and if there's a way