I noticed that it is configurable in job level spark.task.cpus. Anyway to
support on task level?
Thanks.
Zhan Zhang
On Dec 11, 2015, at 10:46 AM, Zhan Zhang wrote:
> Hi Folks,
>
> Is it possible to assign multiple core per task and how? Suppose we have some
> scenario, in which some tasks are really heavy processing each record and
> require multi-threading, and we want to avoid similar tasks assigned to the
> same executors/hosts.
>
> If it is not supported, does it make sense to add this feature. It may seems
> make user worry about more configuration, but by default we can still do 1
> core per task and only advanced users need to be aware of this feature.
>
> Thanks.
>
> Zhan Zhang
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org