Hi, Daniel. I guess that kind of works will start sufficiently in 2.1.0 after PMC's annoucement/reminder on mailing list.
Bests, Dongjoon. On Wednesday, October 26, 2016, Daniel Siegmann < dsiegm...@securityscorecard.io> wrote: > Is the deprecation of JDK 7 and Scala 2.10 documented anywhere outside the > release notes for Spark 2.0.0? I do not consider release notes to be > sufficient public notice for deprecation of supported platforms - this > should be noted in the documentation somewhere. Here are on the only > mentions I could find: > > At http://spark.apache.org/downloads.html it says: > > "*Note: Starting version 2.0, Spark is built with Scala 2.11 by default. > Scala 2.10 users should download the Spark source package and build with > Scala 2.10 support > <http://spark.apache.org/docs/latest/building-spark.html#building-for-scala-210>."* > > At http://spark.apache.org/docs/latest/#downloading it says: > > "Spark runs on Java 7+, Python 2.6+/3.4+ and R 3.1+. For the Scala API, > Spark 2.0.1 uses Scala 2.11. You will need to use a compatible Scala > version (2.11.x)." > > At http://spark.apache.org/docs/latest/programming-guide.html# > linking-with-spark it says: > > - "Spark 2.0.1 is built and distributed to work with Scala 2.11 by > default. (Spark can be built to work with other versions of Scala, too.) To > write applications in Scala, you will need to use a compatible Scala > version (e.g. 2.11.X)." > - "Spark 2.0.1 works with Java 7 and higher. If you are using Java 8, > Spark supports lambda expressions > <http://docs.oracle.com/javase/tutorial/java/javaOO/lambdaexpressions.html> > for concisely writing functions, otherwise you can use the classes in the > org.apache.spark.api.java.function > > <http://spark.apache.org/docs/latest/api/java/index.html?org/apache/spark/api/java/function/package-summary.html> > package." > - "Spark 2.0.1 works with Python 2.6+ or Python 3.4+. It can use the > standard CPython interpreter, so C libraries like NumPy can be used. It > also works with PyPy 2.3+." > > >