Hi Sukant, Regarding the first point: when building spark during my daily work, I always use Scala 2.11 and have only run into build problems once. Assuming a working build I have never had any issues with the resulting artifacts.
More generally however, I would advise you to go with Scala 2.11 under all circumstances. Scala 2.10 has reached end-of-life and, from what I make out of your question, you have the opportunity to switch to a newer technology, so why stay with legacy? Furthermore, Scala 2.12 will be coming out early next year, so I reckon that Spark will switch to Scala 2.11 by default pretty soon*. regards, --Jakob *I'm myself pretty new to the Spark community so please don't take my words on it as gospel On 11 November 2015 at 15:25, Ted Yu <yuzhih...@gmail.com> wrote: > For #1, the published jars are usable. > However, you should build from source for your specific combination of > profiles. > > Cheers > > On Wed, Nov 11, 2015 at 3:22 PM, shajra-cogscale < > sha...@cognitivescale.com> wrote: > >> Hi, >> >> My company isn't using Spark in production yet, but we are using a bit of >> Scala. There's a few people who have wanted to be conservative and keep >> our >> Scala at 2.10 in the event we start using Spark. There are others who >> want >> to move to 2.11 with the idea that by the time we're using Spark it will >> be >> more or less 2.11-ready. >> >> It's hard to make a strong judgement on these kinds of things without >> getting some community feedback. >> >> Looking through the internet I saw: >> >> 1) There's advice to build 2.11 packages from source -- but also published >> jars to Maven Central for 2.11. Are these jars on Maven Central usable >> and >> the advice to build from source outdated? >> >> 2) There's a note that the JDBC RDD isn't 2.11-compliant. This is okay >> for >> us, but is there anything else to worry about? >> >> It would be nice to get some answers to those questions as well as any >> other >> feedback from maintainers or anyone that's used Spark with Scala 2.11 >> beyond >> simple examples. >> >> Thanks, >> Sukant >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Status-of-2-11-support-tp25362.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >