Not possible as of today. See
https://issues.apache.org/jira/browse/SPARK-2387
Hemant Bhanawat
https://www.linkedin.com/in/hemant-bhanawat-92a3811
www.snappydata.io
On Thu, Feb 18, 2016 at 1:19 PM, Shushant Arora
wrote:
> can two stages of single job run in parallel in spark?
>
> e.g one stage is ,map transformation and another is repartition on mapped
> rdd.
>
> rdd.map(function,100).repartition(30);
>
> can it happen that map transformation which is running 100 tasks after few
> of them say (10 ) are finished and spark started another stage repartition
> which started copying data from mapped stage nodes in parallel.
>
> Thanks
>