Most of the time in the PR build is on running tests. How about we also add Scala 2.11 compilation for both main and test without running the tests in the PR build?
Sincerely, DB Tsai ---------------------------------------------------------- Web: https://www.dbtsai.com PGP Key ID: 0x5CED8B896A6BDFA0 On Fri, Nov 16, 2018 at 10:09 PM Marcelo Vanzin <van...@cloudera.com.invalid> wrote: > > Now that the switch to 2.12 by default has been made, it might be good > to have a serious discussion about dropping 2.11 altogether. Many of > the main arguments have already been talked about. But I don't > remember anyone mentioning how easy it would be to break the 2.11 > build now. > > For example, the following works fine in 2.12 but breaks in 2.11: > > java.util.Arrays.asList("hi").stream().forEach(println) > > We had a similar issue when we supported java 1.6 but the builds were > all on 1.7 by default. Every once in a while something would silently > break, because PR builds only check the default. And the jenkins > builds, which are less monitored, would stay broken for a while. > > On Tue, Nov 6, 2018 at 11:13 AM DB Tsai <d_t...@apple.com> wrote: > > > > We made Scala 2.11 as default Scala version in Spark 2.0. Now, the next > > Spark version will be 3.0, so it's a great time to discuss should we make > > Scala 2.12 as default Scala version in Spark 3.0. > > > > Scala 2.11 is EOL, and it came out 4.5 ago; as a result, it's unlikely to > > support JDK 11 in Scala 2.11 unless we're willing to sponsor the needed > > work per discussion in Scala community, > > https://github.com/scala/scala-dev/issues/559#issuecomment-436160166 > > > > We have initial support of Scala 2.12 in Spark 2.4. If we decide to make > > Scala 2.12 as default for Spark 3.0 now, we will have ample time to work on > > bugs and issues that we may run into. > > > > What do you think? > > > > Thanks, > > > > DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, > > Inc > > > > > > --------------------------------------------------------------------- > > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > > > > -- > Marcelo > > --------------------------------------------------------------------- > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org