Thanks Tobias, I have filed a JIRA for it. 2014-09-18 10:09 GMT-07:00 Patrick Wendell <pwend...@gmail.com>:
> I agree, that's a good idea Marcelo. There isn't AFAIK any reason the > client needs to hang there for correct operation. > > On Thu, Sep 18, 2014 at 9:39 AM, Marcelo Vanzin <van...@cloudera.com> > wrote: > > Yes, what Sandy said. > > > > On top of that, I would suggest filing a bug for a new command line > > argument for spark-submit to make the launcher process exit cleanly as > > soon as a cluster job starts successfully. That can be helpful for > > code that launches Spark jobs but monitors the job through different > > means. > > > > On Thu, Sep 18, 2014 at 7:37 AM, Sandy Ryza <sandy.r...@cloudera.com> > wrote: > >> Hi Tobias, > >> > >> YARN cluster mode should have the behavior you're looking for. The > client > >> process will stick around to report on things, but should be able to be > >> killed without affecting the application. If this isn't the behavior > you're > >> observing, and your application isn't failing for a different reason, > >> there's a bug. > >> > >> -Sandy > >> > >> On Thu, Sep 18, 2014 at 10:20 AM, Nicholas Chammas > >> <nicholas.cham...@gmail.com> wrote: > >>> > >>> Dunno about having the application be independent of whether > spark-submit > >>> is still alive, but you can have spark-submit run in a new session in > Linux > >>> using setsid. > >>> > >>> That way even if you terminate your SSH session, spark-submit will keep > >>> running independently. Of course, if you terminate the host running > >>> spark-submit, you will still have problems. > >>> > >>> > >>> On Thu, Sep 18, 2014 at 4:19 AM, Tobias Pfeiffer <t...@preferred.jp> > wrote: > >>>> > >>>> Hi, > >>>> > >>>> I am wondering: Is it possible to run spark-submit in a mode where it > >>>> will start an application on a YARN cluster (i.e., driver and > executors run > >>>> on the cluster) and then forget about it in the sense that the Spark > >>>> application is completely independent from the host that ran the > >>>> spark-submit command and will not be affected if that controlling > machine > >>>> shuts down etc.? I was using spark-submit with YARN in cluster mode, > but > >>>> spark-submit stayed in the foreground and as far as I understood, it > >>>> terminated the application on the cluster when spark-submit was > Ctrl+C'ed. > >>>> > >>>> Thanks > >>>> Tobias > >>> > >>> > >> > > > > > > > > -- > > Marcelo > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > > For additional commands, e-mail: user-h...@spark.apache.org > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >