> On 4 Apr 2016, at 20:58, Ofir Manor <ofir.ma...@equalum.io> wrote:
> 
> I think that a backup plan could be to announce that JDK7 is deprecated in 
> Spark 2.0 and support for it will be fully removed in Spark 2.1. This gives 
> admins enough warning to install JDK8 along side their "main" JDK (or fully 
> migrate to it), while allowing the project to merge JDK8-specific changes to 
> trunk right after the 2.0 release.
> 

Announcing a plan is good; anything which can be done to help mixed JVM 
deployment (documentation, testing) would be useful too

> However, I personally think it is better to drop JDK7 now. I'm sure that both 
> the community and the distributors (Databricks, Cloudera, Hortonworks, MapR, 
> IBM etc) will all rush to help their customers migrate their environment to 
> support Spark 2.0, so I think any backlash won't be dramatic or lasting. 
> 

People using Spark tend to be pretty aggressive about wanting the latest 
version, at least on the 1.x line; so far there've been no major problems 
allowing mixed spark version deployments, provided shared bits of 
infrastructure (spark history server) were recent. Hive metadata repository 
access is the other big issue: moving spark up to hive  1.2.1 addresses that 
for the moment.

I don't know about organisations adoption of JDK8 vs 7; or how anyone would 
react to having to move to java 8 for spark 2. Maybe it'll be a barrier to 
adoption —maybe it'll be an incentive to upgrade.

Oh, I do know that Java 9 is going to be trouble. Different topic.

Reply via email to