I feel good ;) I know there is a lot of interest among Spark users. Since the compiler change won’t force a Spark API change, can we target Spark 2.4?
Sent from my rotary phone. > On Jun 6, 2018, at 11:33 AM, Holden Karau <hol...@pigscanfly.ca> wrote: > > Just chatted with Dean @ the summit and it sounds like from Adriaan there is > a fix in 2.13 for the API change issue that could be back ported to 2.12 so > how about we try and get this ball rolling? > > It sounds like it would also need a closure cleaner change, which could be > backwards compatible but since it’s such a core component and we might want > to be cautious with it, we could when building for 2.11 use the old cleaner > code and for 2.12 use the new code so we don’t break anyone. > > How do folks feel about this? > >> On Sat, Apr 21, 2018 at 5:32 AM Dean Wampler <deanwamp...@gmail.com> wrote: >> Hi, Reynold, >> >> Sorry for the delay in replying; I was traveling. >> >> The Scala changes would avoid the need to change the API now. Basically, the >> compiler would be modified to detect the particular case of the two >> ambiguous, overloaded methods, then pick the best fit in a more >> "intelligent" way. (They can provide more specific details). This would not >> address the closure cleaner changes required. However, the Scala team >> offered to provide suggestions or review changes. >> >> dean >> >> Dean Wampler, Ph.D. >> VP, Fast Data Engineering at Lightbend >> Author: Programming Scala, 2nd Edition, Fast Data Architectures for >> Streaming Applications, and other content from O'Reilly >> @deanwampler >> http://polyglotprogramming.com >> https://github.com/deanwampler >> >>> On Thu, Apr 19, 2018 at 6:46 PM, Reynold Xin <r...@databricks.com> wrote: >>> Forking the thread to focus on Scala 2.12. >>> >>> Dean, >>> >>> There are couple different issues with Scala 2.12 (closure cleaner, API >>> breaking changes). Which one do you think we can address with a Scala >>> upgrade? (The closure cleaner one I haven't spent a lot of time looking at >>> it but it might involve more Spark side changes) >>> >>>> On Thu, Apr 19, 2018 at 3:28 AM, Dean Wampler <deanwamp...@gmail.com> >>>> wrote: >>>> I spoke with Martin Odersky and Lightbend's Scala Team about the known API >>>> issue with method disambiguation. They offered to implement a small patch >>>> in a new release of Scala 2.12 to handle the issue without requiring a >>>> Spark API change. They would cut a 2.12.6 release for it. I'm told that >>>> Scala 2.13 should already handle the issue without modification (it's not >>>> yet released, to be clear). They can also offer feedback on updating the >>>> closure cleaner. >>>> >>>> So, this approach would support Scala 2.12 in Spark, but limited to >>>> 2.12.6+, without the API change requirement, but the closure cleaner would >>>> still need updating. Hence, it could be done for Spark 2.X. >>>> >>>> Let me if you want to pursue this approach. >>>> >>>> dean >>>> >>>> >>>> >>>> Dean Wampler, Ph.D. >>>> VP, Fast Data Engineering at Lightbend >>>> Author: Programming Scala, 2nd Edition, Fast Data Architectures for >>>> Streaming Applications, and other content from O'Reilly >>>> @deanwampler >>>> http://polyglotprogramming.com >>>> https://github.com/deanwampler >>>> >>>>> On Thu, Apr 5, 2018 at 8:13 PM, Marcelo Vanzin <van...@cloudera.com> >>>>> wrote: >>>>> On Thu, Apr 5, 2018 at 10:30 AM, Matei Zaharia <matei.zaha...@gmail.com> >>>>> wrote: >>>>> > Sorry, but just to be clear here, this is the 2.12 API issue: >>>>> > https://issues.apache.org/jira/browse/SPARK-14643, with more details in >>>>> > this doc: >>>>> > https://docs.google.com/document/d/1P_wmH3U356f079AYgSsN53HKixuNdxSEvo8nw_tgLgM/edit. >>>>> > >>>>> > Basically, if we are allowed to change Spark’s API a little to have >>>>> > only one version of methods that are currently overloaded between Java >>>>> > and Scala, we can get away with a single source three for all Scala >>>>> > versions and Java ABI compatibility against any type of Spark (whether >>>>> > using Scala 2.11 or 2.12). >>>>> >>>>> Fair enough. To play devil's advocate, most of those methods seem to >>>>> be marked "Experimental / Evolving", which could be used as a reason >>>>> to change them for this purpose in a minor release. >>>>> >>>>> Not all of them are, though (e.g. foreach / foreachPartition are not >>>>> experimental). >>>>> >>>>> -- >>>>> Marcelo >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>>> >>>> >>> >> > -- > Twitter: https://twitter.com/holdenkarau