Would be good if you can paste the piece of code that you are executing. On Sun, Jun 26, 2016 at 11:21 AM, Ravindra <ravindra.baj...@gmail.com> wrote:
> Hi All, > > May be I need to just set some property or its a known issue. My spark > application hangs in test environment whenever I see following message - > > 16/06/26 11:13:34 INFO DAGScheduler: *Submitting 2 missing tasks from > ShuffleMapStage* 145 (MapPartitionsRDD[590] at rdd at > WriteDataFramesDecorator.scala:61) > 16/06/26 11:13:34 INFO TaskSchedulerImpl: Adding task set 145.0 with 2 > tasks > 16/06/26 11:13:34 INFO TaskSetManager: Starting task 0.0 in stage 145.0 > (TID 186, localhost, PROCESS_LOCAL, 2389 bytes) > 16/06/26 11:13:34 INFO Executor: Running task 0.0 in stage 145.0 (TID 186) > 16/06/26 11:13:34 INFO BlockManager: Found block rdd_575_0 locally > 16/06/26 11:13:34 INFO GenerateMutableProjection: Code generated in 3.796 > ms > 16/06/26 11:13:34 INFO Executor: Finished task 0.0 in stage 145.0 (TID > 186). 2578 bytes result sent to driver > 16/06/26 11:13:34 INFO TaskSetManager: Finished task 0.0 in stage 145.0 > (TID 186) in 24 ms on localhost (1/2) > > It happens with any action. The application works fine whenever I notice > "*Submitting > 1 missing tasks from ShuffleMapStage". *For this I need to tweak the plan > like using repartition, coalesce etc but this also doesn't help always. > > Some of the Spark properties are as given below - > > Name Value > spark.app.id local-1466914377931 > spark.app.name SparkTest > spark.cores.max 3 > spark.default.parallelism 1 > spark.driver.allowMultipleContexts true > spark.executor.id driver > spark.externalBlockStore.folderName > spark-050049bd-c058-4035-bc3d-2e73a08e8d0c > spark.master local[2] > spark.scheduler.mode FIFO > spark.ui.enabled true > > > Thanks, > Ravi. > > -- Cheers!