I am able to use Client.scala or LauncherExecutor.scala as my programmatic 
entry point for Yarn.

Thanks,
Ron

Sent from my iPad

> On Jul 9, 2014, at 7:14 AM, Jerry Lam <chiling...@gmail.com> wrote:
> 
> +1 as well for being able to submit jobs programmatically without using shell 
> script.
> 
> we also experience issues of submitting jobs programmatically without using 
> spark-submit. In fact, even in the Hadoop World, I rarely used "hadoop jar" 
> to submit jobs in shell. 
> 
> 
> 
>> On Wed, Jul 9, 2014 at 9:47 AM, Robert James <srobertja...@gmail.com> wrote:
>> +1 to be able to do anything via SparkConf/SparkContext.  Our app
>> worked fine in Spark 0.9, but, after several days of wrestling with
>> uber jars and spark-submit, and so far failing to get Spark 1.0
>> working, we'd like to go back to doing it ourself with SparkConf.
>> 
>> As the previous poster said, a few scripts should be able to give us
>> the classpath and any other params we need, and be a lot more
>> transparent and debuggable.
>> 
>> On 7/9/14, Surendranauth Hiraman <suren.hira...@velos.io> wrote:
>> > Are there any gaps beyond convenience and code/config separation in using
>> > spark-submit versus SparkConf/SparkContext if you are willing to set your
>> > own config?
>> >
>> > If there are any gaps, +1 on having parity within SparkConf/SparkContext
>> > where possible. In my use case, we launch our jobs programmatically. In
>> > theory, we could shell out to spark-submit but it's not the best option for
>> > us.
>> >
>> > So far, we are only using Standalone Cluster mode, so I'm not knowledgeable
>> > on the complexities of other modes, though.
>> >
>> > -Suren
>> >
>> >
>> >
>> > On Wed, Jul 9, 2014 at 8:20 AM, Koert Kuipers <ko...@tresata.com> wrote:
>> >
>> >> not sure I understand why unifying how you submit app for different
>> >> platforms and dynamic configuration cannot be part of SparkConf and
>> >> SparkContext?
>> >>
>> >> for classpath a simple script similar to "hadoop classpath" that shows
>> >> what needs to be added should be sufficient.
>> >>
>> >> on spark standalone I can launch a program just fine with just SparkConf
>> >> and SparkContext. not on yarn, so the spark-launch script must be doing a
>> >> few things extra there I am missing... which makes things more difficult
>> >> because I am not sure its realistic to expect every application that
>> >> needs
>> >> to run something on spark to be launched using spark-submit.
>> >>  On Jul 9, 2014 3:45 AM, "Patrick Wendell" <pwend...@gmail.com> wrote:
>> >>
>> >>> It fulfills a few different functions. The main one is giving users a
>> >>> way to inject Spark as a runtime dependency separately from their
>> >>> program and make sure they get exactly the right version of Spark. So
>> >>> a user can bundle an application and then use spark-submit to send it
>> >>> to different types of clusters (or using different versions of Spark).
>> >>>
>> >>> It also unifies the way you bundle and submit an app for Yarn, Mesos,
>> >>> etc... this was something that became very fragmented over time before
>> >>> this was added.
>> >>>
>> >>> Another feature is allowing users to set configuration values
>> >>> dynamically rather than compile them inside of their program. That's
>> >>> the one you mention here. You can choose to use this feature or not.
>> >>> If you know your configs are not going to change, then you don't need
>> >>> to set them with spark-submit.
>> >>>
>> >>>
>> >>> On Wed, Jul 9, 2014 at 10:22 AM, Robert James <srobertja...@gmail.com>
>> >>> wrote:
>> >>> > What is the purpose of spark-submit? Does it do anything outside of
>> >>> > the standard val conf = new SparkConf ... val sc = new SparkContext
>> >>> > ... ?
>> >>>
>> >>
>> >
>> >
>> > --
>> >
>> > SUREN HIRAMAN, VP TECHNOLOGY
>> > Velos
>> > Accelerating Machine Learning
>> >
>> > 440 NINTH AVENUE, 11TH FLOOR
>> > NEW YORK, NY 10001
>> > O: (917) 525-2466 ext. 105
>> > F: 646.349.4063
>> > E: suren.hiraman@v <suren.hira...@sociocast.com>elos.io
>> > W: www.velos.io
>> >
> 

Reply via email to