You shouldn't need to do anything special. Are you using a named context?
I'm not sure those work with SparkSqlJob.

By the way, there is a forum on Google groups for the Spark Job Server:
https://groups.google.com/forum/#!forum/spark-jobserver

On Thu, Apr 2, 2015 at 5:10 AM, Harika <matha.har...@gmail.com> wrote:

> Hi,
>
> I am trying to Spark Jobserver(
> https://github.com/spark-jobserver/spark-jobserver
> <https://github.com/spark-jobserver/spark-jobserver>  ) for running Spark
> SQL jobs.
>
> I was able to start the server but when I run my application(my Scala class
> which extends SparkSqlJob), I am getting the following as response:
>
> {
>   "status": "ERROR",
>   "result": "Invalid job type for this context"
> }
>
> Can any one suggest me what is going wrong or provide a detailed procedure
> for setting up jobserver for SparkSQL?
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Setup-Spark-jobserver-for-Spark-SQL-tp22352.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to