I think that's fine, personally. Anyone using JDK 8 should / probably
is on a recent release.

On Thu, Oct 24, 2019 at 8:56 PM Dongjoon Hyun <dongjoon.h...@gmail.com> wrote:
>
> Thank you for reply, Sean, Shane, Takeshi.
>
> The reason is that there is a PR to aim to add `-XX:OnOutOfMemoryError="kill 
> -9 %p"` as a default behavior at 3.0.0.
> (Please note that the PR will add it by *default* always. There is no way for 
> user to remove it.)
>
>     - [SPARK-27900][CORE][K8s] Add `spark.driver.killOnOOMError` flag in 
> cluster mode
>     - https://github.com/apache/spark/pull/26161
>
> If we can deprecate old JDK8 versions, we are able to use JVM option 
> `ExitOnOutOfMemoryError` instead.
> (This is added at JDK 8u92. In my previous email, 8u82 was a typo.)
>
>     - 
> https://www.oracle.com/technetwork/java/javase/8u92-relnotes-2949471.html
>
> All versions of JDK8 are not the same naturally. For example, Hadoop 
> community also have the following document although they are not specifying 
> the minimum versions.
>
>     - https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+Java+Versions
>
> Bests,
> Dongjoon.
>
>
> On Thu, Oct 24, 2019 at 6:05 PM Takeshi Yamamuro <linguin....@gmail.com> 
> wrote:
>>
>> Hi, Dongjoon
>>
>> It might be worth clearly describing which jdk versions we check in the 
>> testing infra
>> in some documents, e.g., https://spark.apache.org/docs/latest/#downloading
>>
>> btw, any other project announcing the minimum support jdk version?
>> It seems that hadoop does not.
>>
>> On Fri, Oct 25, 2019 at 6:51 AM Sean Owen <sro...@gmail.com> wrote:
>>>
>>> Probably, but what is the difference that makes it different to
>>> support u81 vs later?
>>>
>>> On Thu, Oct 24, 2019 at 4:39 PM Dongjoon Hyun <dongjoon.h...@gmail.com> 
>>> wrote:
>>> >
>>> > Hi, All.
>>> >
>>> > Apache Spark 3.x will support both JDK8 and JDK11.
>>> >
>>> > I'm wondering if we can have a minimum JDK8 version in Apache Spark 3.0.
>>> >
>>> > Specifically, can we start to deprecate JDK8u81 and older at 3.0.
>>> >
>>> > Currently, Apache Spark testing infra are testing only with jdk1.8.0_191 
>>> > and above.
>>> >
>>> > Bests,
>>> > Dongjoon.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>
>>
>>
>> --
>> ---
>> Takeshi Yamamuro

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to