Hi, When running some tests on EC2 with spark, I notice that: the tasks are not fairly distributed to executor.
For example, a map action produces 4 tasks, but they all go to the Executors (3) - *Memory:* 0.0 B Used (19.0 GB Total) - *Disk:* 0.0 B Used Executor IDAddressRDD blocksMemory usedDisk usedActive tasksFailed tasksComplete tasksTotal tasks0ip-10-10-141-143.ec2.internal:5281600.0 B / 6.3 GB0.0 B4004 1ip-10-40-38-190.ec2.internal:6031400.0 B / 6.3 GB0.0 B00002 ip-10-62-35-223.ec2.internal:4050000.0 B / 6.3 GB0.0 B0000
