Thank you very much, well documented. Thanks Sri
On Wed, Jan 27, 2016 at 8:46 PM, Deenar Toraskar <deenar.toras...@gmail.com> wrote: > Sri > > Look at the instructions here. They are for 1.5.1, but should also work > for 1.6 > > > https://www.linkedin.com/pulse/running-spark-151-cdh-deenar-toraskar-cfa?trk=hp-feed-article-title-publish&trkSplashRedir=true&forceNoSplash=true > > Deenar > > > On 27 January 2016 at 20:16, Koert Kuipers <ko...@tresata.com> wrote: > >> you need to build spark 1.6 for your hadoop distro, and put that on the >> proxy node and configure it correctly to find your cluster (hdfs and yarn). >> then use the spark-submit script for that spark 1.6 version to launch your >> application on yarn >> >> On Wed, Jan 27, 2016 at 3:11 PM, sri hari kali charan Tummala < >> kali.tumm...@gmail.com> wrote: >> >>> Hi Koert, >>> >>> I am submitting my code (spark jar ) using spark-submit in proxy node , >>> I checked the version of the cluster and node its says 1.2 I dint really >>> understand what you mean. >>> >>> can I ask yarn to use different version of spark ? or should I say >>> override the spark_home variables to look at 1.6 spark jar ? >>> >>> Thanks >>> Sri >>> >>> On Wed, Jan 27, 2016 at 7:45 PM, Koert Kuipers <ko...@tresata.com> >>> wrote: >>> >>>> If you have yarn you can just launch your spark 1.6 job from a single >>>> machine with spark 1.6 available on it and ignore the version of spark >>>> (1.2) that is installed >>>> On Jan 27, 2016 11:29, "kali.tumm...@gmail.com" <kali.tumm...@gmail.com> >>>> wrote: >>>> >>>>> Hi All, >>>>> >>>>> Just realized cloudera version of spark on my cluster is 1.2, the jar >>>>> which >>>>> I built using maven is version 1.6 which is causing issue. >>>>> >>>>> Is there a way to run spark version 1.6 in 1.2 version of spark ? >>>>> >>>>> Thanks >>>>> Sri >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/how-to-run-latest-version-of-spark-in-old-version-of-spark-in-cloudera-cluster-tp26087.html >>>>> Sent from the Apache Spark User List mailing list archive at >>>>> Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>> >>>>> >>> >>> >>> -- >>> Thanks & Regards >>> Sri Tummala >>> >>> >> > -- Thanks & Regards Sri Tummala