something like this import org.apache.spark.TaskContext ds.map(r => { val taskContext = TaskContext.get() if (taskContext.partitionId == 1000) { throw new RuntimeException } r })
On Mon, Feb 11, 2019 at 8:41 AM Serega Sheypak <serega.shey...@gmail.com> wrote: > > I need to crash task which does repartition. > > пн, 11 февр. 2019 г. в 10:37, Gabor Somogyi <gabor.g.somo...@gmail.com>: >> >> What blocks you to put if conditions inside the mentioned map function? >> >> On Mon, Feb 11, 2019 at 10:31 AM Serega Sheypak <serega.shey...@gmail.com> >> wrote: >>> >>> Yeah, but I don't need to crash entire app, I want to fail several tasks or >>> executors and then wait for completion. >>> >>> вс, 10 февр. 2019 г. в 21:49, Gabor Somogyi <gabor.g.somo...@gmail.com>: >>>> >>>> Another approach is adding artificial exception into the application's >>>> source code like this: >>>> >>>> val query = input.toDS.map(_ / 0).writeStream.format("console").start() >>>> >>>> G >>>> >>>> >>>> On Sun, Feb 10, 2019 at 9:36 PM Serega Sheypak <serega.shey...@gmail.com> >>>> wrote: >>>>> >>>>> Hi BR, >>>>> thanks for your reply. I want to mimic the issue and kill tasks at a >>>>> certain stage. Killing executor is also an option for me. >>>>> I'm curious how do core spark contributors test spark fault tolerance? >>>>> >>>>> >>>>> вс, 10 февр. 2019 г. в 16:57, Gabor Somogyi <gabor.g.somo...@gmail.com>: >>>>>> >>>>>> Hi Serega, >>>>>> >>>>>> If I understand your problem correctly you would like to kill one >>>>>> executor only and the rest of the app has to be untouched. >>>>>> If that's true yarn -kill is not what you want because it stops the >>>>>> whole application. >>>>>> >>>>>> I've done similar thing when tested/testing Spark's HA features. >>>>>> - jps -vlm | grep >>>>>> "org.apache.spark.executor.CoarseGrainedExecutorBackend.*applicationid" >>>>>> - kill -9 pidofoneexecutor >>>>>> >>>>>> Be aware if it's a multi-node cluster check whether at least one process >>>>>> runs on a specific node(it's not required). >>>>>> Happy killing... >>>>>> >>>>>> BR, >>>>>> G >>>>>> >>>>>> >>>>>> On Sun, Feb 10, 2019 at 4:19 PM Jörn Franke <jornfra...@gmail.com> wrote: >>>>>>> >>>>>>> yarn application -kill applicationid ? >>>>>>> >>>>>>> > Am 10.02.2019 um 13:30 schrieb Serega Sheypak >>>>>>> > <serega.shey...@gmail.com>: >>>>>>> > >>>>>>> > Hi there! >>>>>>> > I have weird issue that appears only when tasks fail at specific >>>>>>> > stage. I would like to imitate failure on my own. >>>>>>> > The plan is to run problematic app and then kill entire executor or >>>>>>> > some tasks when execution reaches certain stage. >>>>>>> > >>>>>>> > Is it do-able? >>>>>>> >>>>>>> --------------------------------------------------------------------- >>>>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >>>>>>> -- Sent from my iPhone --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org