Thanks Tobias, I have filed a JIRA for it.

2014-09-18 10:09 GMT-07:00 Patrick Wendell <pwend...@gmail.com>:

> I agree, that's a good idea Marcelo. There isn't AFAIK any reason the
> client needs to hang there for correct operation.
>
> On Thu, Sep 18, 2014 at 9:39 AM, Marcelo Vanzin <van...@cloudera.com>
> wrote:
> > Yes, what Sandy said.
> >
> > On top of that, I would suggest filing a bug for a new command line
> > argument for spark-submit to make the launcher process exit cleanly as
> > soon as a cluster job starts successfully. That can be helpful for
> > code that launches Spark jobs but monitors the job through different
> > means.
> >
> > On Thu, Sep 18, 2014 at 7:37 AM, Sandy Ryza <sandy.r...@cloudera.com>
> wrote:
> >> Hi Tobias,
> >>
> >> YARN cluster mode should have the behavior you're looking for.  The
> client
> >> process will stick around to report on things, but should be able to be
> >> killed without affecting the application.  If this isn't the behavior
> you're
> >> observing, and your application isn't failing for a different reason,
> >> there's a bug.
> >>
> >> -Sandy
> >>
> >> On Thu, Sep 18, 2014 at 10:20 AM, Nicholas Chammas
> >> <nicholas.cham...@gmail.com> wrote:
> >>>
> >>> Dunno about having the application be independent of whether
> spark-submit
> >>> is still alive, but you can have spark-submit run in a new session in
> Linux
> >>> using setsid.
> >>>
> >>> That way even if you terminate your SSH session, spark-submit will keep
> >>> running independently. Of course, if you terminate the host running
> >>> spark-submit, you will still have problems.
> >>>
> >>>
> >>> On Thu, Sep 18, 2014 at 4:19 AM, Tobias Pfeiffer <t...@preferred.jp>
> wrote:
> >>>>
> >>>> Hi,
> >>>>
> >>>> I am wondering: Is it possible to run spark-submit in a mode where it
> >>>> will start an application on a YARN cluster (i.e., driver and
> executors run
> >>>> on the cluster) and then forget about it in the sense that the Spark
> >>>> application is completely independent from the host that ran the
> >>>> spark-submit command and will not be affected if that controlling
> machine
> >>>> shuts down etc.? I was using spark-submit with YARN in cluster mode,
> but
> >>>> spark-submit stayed in the foreground and as far as I understood, it
> >>>> terminated the application on the cluster when spark-submit was
> Ctrl+C'ed.
> >>>>
> >>>> Thanks
> >>>> Tobias
> >>>
> >>>
> >>
> >
> >
> >
> > --
> > Marcelo
> >
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> > For additional commands, e-mail: user-h...@spark.apache.org
> >
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to