Hi, Dongjoon It might be worth clearly describing which jdk versions we check in the testing infra in some documents, e.g., https://spark.apache.org/docs/latest/#downloading
btw, any other project announcing the minimum support jdk version? It seems that hadoop does not. On Fri, Oct 25, 2019 at 6:51 AM Sean Owen <sro...@gmail.com> wrote: > Probably, but what is the difference that makes it different to > support u81 vs later? > > On Thu, Oct 24, 2019 at 4:39 PM Dongjoon Hyun <dongjoon.h...@gmail.com> > wrote: > > > > Hi, All. > > > > Apache Spark 3.x will support both JDK8 and JDK11. > > > > I'm wondering if we can have a minimum JDK8 version in Apache Spark 3.0. > > > > Specifically, can we start to deprecate JDK8u81 and older at 3.0. > > > > Currently, Apache Spark testing infra are testing only with jdk1.8.0_191 > and above. > > > > Bests, > > Dongjoon. > > --------------------------------------------------------------------- > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > -- --- Takeshi Yamamuro