And for the record, the issue is here: https://issues.apache.org/jira/browse/SPARK-3591
On Thu, Sep 18, 2014 at 1:19 PM, Andrew Or <and...@databricks.com> wrote: > Thanks Tobias, I have filed a JIRA for it. > > 2014-09-18 10:09 GMT-07:00 Patrick Wendell <pwend...@gmail.com>: > > I agree, that's a good idea Marcelo. There isn't AFAIK any reason the >> client needs to hang there for correct operation. >> >> On Thu, Sep 18, 2014 at 9:39 AM, Marcelo Vanzin <van...@cloudera.com> >> wrote: >> > Yes, what Sandy said. >> > >> > On top of that, I would suggest filing a bug for a new command line >> > argument for spark-submit to make the launcher process exit cleanly as >> > soon as a cluster job starts successfully. That can be helpful for >> > code that launches Spark jobs but monitors the job through different >> > means. >> > >> > On Thu, Sep 18, 2014 at 7:37 AM, Sandy Ryza <sandy.r...@cloudera.com> >> wrote: >> >> Hi Tobias, >> >> >> >> YARN cluster mode should have the behavior you're looking for. The >> client >> >> process will stick around to report on things, but should be able to be >> >> killed without affecting the application. If this isn't the behavior >> you're >> >> observing, and your application isn't failing for a different reason, >> >> there's a bug. >> >> >> >> -Sandy >> >> >> >> On Thu, Sep 18, 2014 at 10:20 AM, Nicholas Chammas >> >> <nicholas.cham...@gmail.com> wrote: >> >>> >> >>> Dunno about having the application be independent of whether >> spark-submit >> >>> is still alive, but you can have spark-submit run in a new session in >> Linux >> >>> using setsid. >> >>> >> >>> That way even if you terminate your SSH session, spark-submit will >> keep >> >>> running independently. Of course, if you terminate the host running >> >>> spark-submit, you will still have problems. >> >>> >> >>> >> >>> On Thu, Sep 18, 2014 at 4:19 AM, Tobias Pfeiffer <t...@preferred.jp> >> wrote: >> >>>> >> >>>> Hi, >> >>>> >> >>>> I am wondering: Is it possible to run spark-submit in a mode where it >> >>>> will start an application on a YARN cluster (i.e., driver and >> executors run >> >>>> on the cluster) and then forget about it in the sense that the Spark >> >>>> application is completely independent from the host that ran the >> >>>> spark-submit command and will not be affected if that controlling >> machine >> >>>> shuts down etc.? I was using spark-submit with YARN in cluster mode, >> but >> >>>> spark-submit stayed in the foreground and as far as I understood, it >> >>>> terminated the application on the cluster when spark-submit was >> Ctrl+C'ed. >> >>>> >> >>>> Thanks >> >>>> Tobias >> >>> >> >>> >> >> >> > >> > >> > >> > -- >> > Marcelo >> > >> > --------------------------------------------------------------------- >> > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> > For additional commands, e-mail: user-h...@spark.apache.org >> > >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >