spark-submit --class “Classname"   --master yarn-cluster
jarfile(withcomplete path)

This should work.

On Fri, Oct 10, 2014 at 8:36 AM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> Yes, you can run it with --master=spark://your-spark-uri:7077 i believe.
>
> Thanks
> Best Regards
>
> On Fri, Oct 10, 2014 at 7:03 PM, Theodore Si <sjyz...@gmail.com> wrote:
>
>> Should I pack the example into a jar file and submit it on master?
>>
>> On Fri, Oct 10, 2014 at 9:32 PM, Theodore Si <sjyz...@gmail.com> wrote:
>>
>>> But I cannot do this via using
>>>
>>> ./bin/run-example SparkPi 10
>>>
>>> right?
>>>
>>>
>>> On Fri, Oct 10, 2014 at 6:04 PM, Akhil Das <ak...@sigmoidanalytics.com>
>>> wrote:
>>>
>>>> This is how the spark-cluster looks like
>>>>
>>>>
>>>> So your driver program (example application) can be ran on the master
>>>> (or anywhere which has access to the master - clustermanager) and the
>>>> workers will execute it.
>>>>
>>>> Thanks
>>>> Best Regards
>>>>
>>>> On Fri, Oct 10, 2014 at 2:47 PM, Theodore Si <sjyz...@gmail.com> wrote:
>>>>
>>>>> Hi all,
>>>>>
>>>>> I want to use two nodes for test, one as master, the other worker.
>>>>> Can I submit the example application included in Spark source code
>>>>> tarball on master to let it run on the worker?
>>>>> What should I do?
>>>>>
>>>>> BR,
>>>>> Theo
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>>
>>>>>
>>>>
>>>
>>
>


-- 
Regards,
Haripriya Ayyalasomayajula

Reply via email to