Thank you. I created a PR for that. For now, the minimum requirement is 8u92 in that PR.
https://github.com/apache/spark/pull/26249 Bests, Dongjoon. On Thu, Oct 24, 2019 at 7:55 PM Sean Owen <sro...@gmail.com> wrote: > I think that's fine, personally. Anyone using JDK 8 should / probably > is on a recent release. > > On Thu, Oct 24, 2019 at 8:56 PM Dongjoon Hyun <dongjoon.h...@gmail.com> > wrote: > > > > Thank you for reply, Sean, Shane, Takeshi. > > > > The reason is that there is a PR to aim to add > `-XX:OnOutOfMemoryError="kill -9 %p"` as a default behavior at 3.0.0. > > (Please note that the PR will add it by *default* always. There is no > way for user to remove it.) > > > > - [SPARK-27900][CORE][K8s] Add `spark.driver.killOnOOMError` flag in > cluster mode > > - https://github.com/apache/spark/pull/26161 > > > > If we can deprecate old JDK8 versions, we are able to use JVM option > `ExitOnOutOfMemoryError` instead. > > (This is added at JDK 8u92. In my previous email, 8u82 was a typo.) > > > > - > https://www.oracle.com/technetwork/java/javase/8u92-relnotes-2949471.html > > > > All versions of JDK8 are not the same naturally. For example, Hadoop > community also have the following document although they are not specifying > the minimum versions. > > > > - > https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+Java+Versions > > > > Bests, > > Dongjoon. > > > > > > On Thu, Oct 24, 2019 at 6:05 PM Takeshi Yamamuro <linguin....@gmail.com> > wrote: > >> > >> Hi, Dongjoon > >> > >> It might be worth clearly describing which jdk versions we check in the > testing infra > >> in some documents, e.g., > https://spark.apache.org/docs/latest/#downloading > >> > >> btw, any other project announcing the minimum support jdk version? > >> It seems that hadoop does not. > >> > >> On Fri, Oct 25, 2019 at 6:51 AM Sean Owen <sro...@gmail.com> wrote: > >>> > >>> Probably, but what is the difference that makes it different to > >>> support u81 vs later? > >>> > >>> On Thu, Oct 24, 2019 at 4:39 PM Dongjoon Hyun <dongjoon.h...@gmail.com> > wrote: > >>> > > >>> > Hi, All. > >>> > > >>> > Apache Spark 3.x will support both JDK8 and JDK11. > >>> > > >>> > I'm wondering if we can have a minimum JDK8 version in Apache Spark > 3.0. > >>> > > >>> > Specifically, can we start to deprecate JDK8u81 and older at 3.0. > >>> > > >>> > Currently, Apache Spark testing infra are testing only with > jdk1.8.0_191 and above. > >>> > > >>> > Bests, > >>> > Dongjoon. > >>> > >>> --------------------------------------------------------------------- > >>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > >>> > >> > >> > >> -- > >> --- > >> Takeshi Yamamuro >