Not possible as of today. See
https://issues.apache.org/jira/browse/SPARK-2387
Hemant Bhanawat
https://www.linkedin.com/in/hemant-bhanawat-92a3811
www.snappydata.io
On Thu, Feb 18, 2016 at 1:19 PM, Shushant Arora
wrote:
> can two stages of single job run in parallel
can two stages of single job run in parallel in spark?
e.g one stage is ,map transformation and another is repartition on mapped
rdd.
rdd.map(function,100).repartition(30);
can it happen that map transformation which is running 100 tasks after few
of them say (10 ) are finished and spark