for a shuffle action, I see the data processed by the cores of the same executor is not balanced and of course the one takes the longest time will slow down the whole process time.
So I would like to know if it is possible to make some modification, so the data will be shared equally between the cores.
I use Spark 2.4 on aws emr and s3.