Can you file a JIRA with the detailed steps to reproduce the problem? On Fri, May 29, 2015 at 2:59 AM, Alex Nakos <ana...@gmail.com> wrote:
> Hi- > > I’ve just built the latest spark RC from source (1.4.0 RC3) and can > confirm that the spark shell is still NOT working properly on 2.11. No > classes in the jar I've specified with the —jars argument on the command > line are available in the REPL. > > > Cheers > Alex > > On Thu, May 28, 2015 at 8:38 AM, Tathagata Das <t...@databricks.com> > wrote: > >> Would be great if you guys can test out the Spark 1.4.0 RC2 (RC3 coming >> out soon) with Scala 2.11 and report issues. >> >> TD >> >> On Tue, May 26, 2015 at 9:15 AM, Koert Kuipers <ko...@tresata.com> wrote: >> >>> we are still running into issues with spark-shell not working on 2.11, >>> but we are running on somewhat older master so maybe that has been resolved >>> already. >>> >>> On Tue, May 26, 2015 at 11:48 AM, Dean Wampler <deanwamp...@gmail.com> >>> wrote: >>> >>>> Most of the 2.11 issues are being resolved in Spark 1.4. For a while, >>>> the Spark project has published maven artifacts that are compiled with 2.11 >>>> and 2.10, although the downloads at >>>> http://spark.apache.org/downloads.html are still all for 2.10. >>>> >>>> Dean Wampler, Ph.D. >>>> Author: Programming Scala, 2nd Edition >>>> <http://shop.oreilly.com/product/0636920033073.do> (O'Reilly) >>>> Typesafe <http://typesafe.com> >>>> @deanwampler <http://twitter.com/deanwampler> >>>> http://polyglotprogramming.com >>>> >>>> On Tue, May 26, 2015 at 10:33 AM, Ritesh Kumar Singh < >>>> riteshoneinamill...@gmail.com> wrote: >>>> >>>>> Yes, recommended version is 2.10 as all the features are not supported >>>>> by 2.11 version. Kafka libraries and JDBC components are yet to be ported >>>>> to 2.11 version. And so if your project doesn't depend on these >>>>> components, >>>>> you can give v2.11 a try. >>>>> >>>>> Here's a link >>>>> <https://spark.apache.org/docs/1.2.0/building-spark.html#building-for-scala-211> >>>>> for >>>>> building with 2.11 version. >>>>> >>>>> Though, you won't be running into any issues if you try v2.10 as of >>>>> now. But then again, the future releases will have to shift to 2.11 >>>>> version >>>>> once support for v2.10 ends in the long run. >>>>> >>>>> >>>>> On Tue, May 26, 2015 at 8:21 PM, Punyashloka Biswal < >>>>> punya.bis...@gmail.com> wrote: >>>>> >>>>>> Dear Spark developers and users, >>>>>> >>>>>> Am I correct in believing that the recommended version of Scala to >>>>>> use with Spark is currently 2.10? Is there any plan to switch to 2.11 in >>>>>> future? Are there any advantages to using 2.11 today? >>>>>> >>>>>> Regards, >>>>>> Punya >>>>> >>>>> >>>>> >>>> >>> >> >