Thank you all for your responses. Will try them out.

On Thu, Sep 3, 2020 at 12:06 AM tianlangstudio <tianlangstu...@aliyun.com>
wrote:

> Hello, Eric
> Maybe you can use  Spark JobServer 0.10.0
> https://github.com/spark-jobserver/spark-jobserverl
> We used this with Spark 1.6, and it is awesome. You know
> the project is still very active. So highly recommend it to you
>
>    <https://www.upwork.com/fl/huanqingzhu>
> <https://www.tianlang.tech/>Fusion Zhu <https://www.tianlang.tech/>
>
> ------------------------------------------------------------------
> 发件人:Eric Beabes <mailinglist...@gmail.com>
> 发送时间:2020年9月3日(星期四) 04:58
> 收件人:spark-user <user@spark.apache.org>
> 主 题:Submitting Spark Job thru REST API?
>
> Under Spark 2.4 is it possible to submit a Spark job thru REST API - just
> like the Flink job?
>
> Here's the use case: We need to submit a Spark Job to the EMR cluster but
> our security team is not allowing us to submit a job from the Master node
> or thru UI. They want us to create a "Docker Container" to submit a job.
>
> If it's possible to submit the Spark job thru REST then we don't need to
> install Spark/Hadoop JARs on the Container. If it's not possible to use
> REST API, can we do something like this?
>
> spark-2.4.6-bin-hadoop2.7/bin/spark-submit \
>  --class myclass --master "yarn url" --deploy-mode cluster \
>
> In other words, instead of --master yarn, specify a URL. Would this still
> work the same way?
>
>

Reply via email to