What blocks you to put if conditions inside the mentioned map function? On Mon, Feb 11, 2019 at 10:31 AM Serega Sheypak <serega.shey...@gmail.com> wrote:
> Yeah, but I don't need to crash entire app, I want to fail several tasks > or executors and then wait for completion. > > вс, 10 февр. 2019 г. в 21:49, Gabor Somogyi <gabor.g.somo...@gmail.com>: > >> Another approach is adding artificial exception into the application's >> source code like this: >> >> val query = input.toDS.map(_ / 0).writeStream.format("console").start() >> >> G >> >> >> On Sun, Feb 10, 2019 at 9:36 PM Serega Sheypak <serega.shey...@gmail.com> >> wrote: >> >>> Hi BR, >>> thanks for your reply. I want to mimic the issue and kill tasks at a >>> certain stage. Killing executor is also an option for me. >>> I'm curious how do core spark contributors test spark fault tolerance? >>> >>> >>> вс, 10 февр. 2019 г. в 16:57, Gabor Somogyi <gabor.g.somo...@gmail.com>: >>> >>>> Hi Serega, >>>> >>>> If I understand your problem correctly you would like to kill one >>>> executor only and the rest of the app has to be untouched. >>>> If that's true yarn -kill is not what you want because it stops the >>>> whole application. >>>> >>>> I've done similar thing when tested/testing Spark's HA features. >>>> - jps -vlm | grep >>>> "org.apache.spark.executor.CoarseGrainedExecutorBackend.*applicationid" >>>> - kill -9 pidofoneexecutor >>>> >>>> Be aware if it's a multi-node cluster check whether at least one >>>> process runs on a specific node(it's not required). >>>> Happy killing... >>>> >>>> BR, >>>> G >>>> >>>> >>>> On Sun, Feb 10, 2019 at 4:19 PM Jörn Franke <jornfra...@gmail.com> >>>> wrote: >>>> >>>>> yarn application -kill applicationid ? >>>>> >>>>> > Am 10.02.2019 um 13:30 schrieb Serega Sheypak < >>>>> serega.shey...@gmail.com>: >>>>> > >>>>> > Hi there! >>>>> > I have weird issue that appears only when tasks fail at specific >>>>> stage. I would like to imitate failure on my own. >>>>> > The plan is to run problematic app and then kill entire executor or >>>>> some tasks when execution reaches certain stage. >>>>> > >>>>> > Is it do-able? >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: user-unsubscr...@spark.apache.org >>>>> >>>>>