My take on the 2-3 tasks per CPU core is that you want to ensure you are
utilizing the cores to the max, which means it will help you with scaling
and performance. The question would be why not 1 task per core? The reason
is that you can probably get a good handle on the average execution time
per
Dear spark users,
>From this site https://spark.apache.org/docs/latest/tuning.html where it
offers recommendation on setting the level of parallelism
Clusters will not be fully utilized unless you set the level of parallelism
> for each operation high enough. Spark automatically sets the number