You cannot run Spark in cluster mode by instantiating a SparkContext like
that.

You have to launch it with the "spark-submit" command line script.

On Thu, Jul 9, 2015 at 2:23 PM, jegordon <jgordo...@gmail.com> wrote:

> Hi to all,
>
> Is there any way to run pyspark scripts with yarn-cluster mode without
> using
> the spark-submit script? I need it in this way because i will integrate
> this
> code into a django web app.
>
> When i try to run any script in yarn-cluster mode i got the following error
> :
>
> org.apache.spark.SparkException: Detected yarn-cluster mode, but isn't
> running on a cluster. Deployment to YARN is not supported directly by
> SparkContext. Please use spark-submit.
>
>
> I'm creating the sparkContext in the following way :
>
>         conf = (SparkConf()
>             .setMaster("yarn-cluster")
>             .setAppName("DataFrameTest"))
>
>         sc = SparkContext(conf = conf)
>
>         #Dataframe code ....
>
> Thanks
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Pyspark-not-working-on-yarn-cluster-mode-tp23755.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>


-- 
Marcelo

Reply via email to