Ok, got it -- it's really just an argument for not all of 2.11, 2.12 and 2.13 at the same time; always 2.12; now figure out when we stop 2.11 support and start 2.13 support.
On Wed, Nov 7, 2018 at 11:10 AM Sean Owen <sro...@gmail.com> wrote: > It's not making 2.12 the default, but not dropping 2.11. Supporting > 2.13 could mean supporting 3 Scala versions at once, which I claim is > just too much. I think the options are likely: > > - Support 2.11, 2.12 in Spark 3.0. Deprecate 2.11 and make 2.12 the > default. Add 2.13 support in 3.x and drop 2.11 in the same release > - Deprecate 2.11 right now via announcement and/or Spark 2.4.1 soon. > Drop 2.11 support in Spark 3.0, and support only 2.12. > - (same as above, but add Spark 2.13 support if possible for Spark 3.0) > > > On Wed, Nov 7, 2018 at 12:32 PM Mark Hamstra <m...@clearstorydata.com> > wrote: > > > > I'm not following "exclude Scala 2.13". Is there something inherent in > making 2.12 the default Scala version in Spark 3.0 that would prevent us > from supporting the option of building with 2.13? > > > > On Tue, Nov 6, 2018 at 5:48 PM Sean Owen <sro...@gmail.com> wrote: > >> > >> That's possible here, sure. The issue is: would you exclude Scala 2.13 > >> support in 3.0 for this, if it were otherwise ready to go? > >> I think it's not a hard rule that something has to be deprecated > >> previously to be removed in a major release. The notice is helpful, > >> sure, but there are lots of ways to provide that notice to end users. > >> Lots of things are breaking changes in a major release. Or: deprecate > >> in Spark 2.4.1, if desired? > >> > >> On Tue, Nov 6, 2018 at 7:36 PM Wenchen Fan <cloud0...@gmail.com> wrote: > >> > > >> > We make Scala 2.11 the default one in Spark 2.0, then drop Scala 2.10 > in Spark 2.3. Shall we follow it and drop Scala 2.11 at some point of Spark > 3.x? > >> > > >> > On Wed, Nov 7, 2018 at 8:55 AM Reynold Xin <r...@databricks.com> > wrote: > >> >> > >> >> Have we deprecated Scala 2.11 already in an existing release? > >> > >> --------------------------------------------------------------------- > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > >> >