Thank you Dongjoon. Please check out the latest code from test-spark-jdk11 <https://github.com/wangyum/test-spark-jdk11>. It works with JDK 1.8. One workaround is to install the Spark packages to local maven repository using hadoop-3.2 profile and JDK 1.8.
On Mon, Oct 28, 2019 at 5:03 AM Dongjoon Hyun <dongjoon.h...@gmail.com> wrote: > Hi, Yuming. > > Is the project working correctly on JDK8 with you? > > When I simply cloned your repo and did `mvn clean package` on > JDK 1.8.0_232, it seems not to pass the UTs. > > I also tried to rerun after ignoring two ORC table test like the > followings, but the UT is failing. > > ~/A/test-spark-jdk11:master$ git diff | grep 'ORC table' > - test("Datasource ORC table") { > + ignore("Datasource ORC table") { > - test("Hive ORC table") { > + ignore("Hive ORC table") { > > ~/A/test-spark-jdk11:master$ mvn clean package > ... > - Hive ORC table !!! IGNORED !!! > Run completed in 36 seconds, 999 milliseconds. > Total number of tests run: 2 > Suites: completed 3, aborted 0 > Tests: succeeded 1, failed 1, canceled 0, ignored 2, pending 0 > *** 1 TEST FAILED *** > > ~/A/test-spark-jdk11:master$ java -version > openjdk version "1.8.0_232" > OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_232-b09) > OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.232-b09, mixed mode) > > > Bests, > Dongjoon. > > On Sun, Oct 27, 2019 at 1:38 PM Dongjoon Hyun <dongjoon.h...@gmail.com> > wrote: > >> It seems not a Hadoop issue, doesn't it? >> >> What Yuming pointed seems to be `Hive 2.3.6` profile implementation issue >> which is enabled only when `Hadoop 3.2`. >> >> From my side, I'm +1 for publishing jars which depends on `Hadoop 3.2.0 / >> Hive 2.3.6` jars to Maven since Apache Spark 3.0.0. >> >> For the others, I'd like to mention that this implies the followings, too. >> >> 1. We are not going to use Hive 1.2.1 library. Only Hadoop-2.7 profile >> tarball distribution will use Hive 1.2.1. >> 2. Although we depends on Hadoop 3.2.0, Hadoop 3.2.1 changes their Guava >> library version significantly. >> So, it requires some attentions in Apache Spark. Otherwise, we may >> hit some issues on Hadoop 3.2.1+ runtime later. >> >> Thanks, >> Dongjoon. >> >> >> On Sun, Oct 27, 2019 at 7:31 AM Sean Owen <sro...@gmail.com> wrote: >> >>> Is the Spark artifact actually any different between those builds? I >>> thought it just affected what else was included in the binary tarball. >>> If it matters, yes I'd publish a "Hadoop 3" version to Maven. (Scala >>> 2.12 is the only supported Scala version). >>> >>> On Sun, Oct 27, 2019 at 4:35 AM Yuming Wang <wgy...@gmail.com> wrote: >>> > >>> > Do we need to publish the Scala 2.12 + hadoop 3.2 jar packages to the >>> Maven repository? Otherwise it will throw a NoSuchMethodError on Java 11. >>> > Here is an example: >>> > >>> https://github.com/wangyum/test-spark-jdk11/blob/master/src/test/scala/test/spark/HiveTableSuite.scala#L34-L38 >>> > >>> https://github.com/wangyum/test-spark-jdk11/commit/927ce7d3766881fba98f2434055fa3a1d1544ad2/checks?check_suite_id=283076578 >>> > >>> > >>> > On Sat, Oct 26, 2019 at 10:41 AM Takeshi Yamamuro < >>> linguin....@gmail.com> wrote: >>> >> >>> >> Thanks for that work! >>> >> >>> >> > I don't think JDK 11 is a separate release (by design). We build >>> >> > everything targeting JDK 8 and it should work on JDK 11 too. >>> >> +1. a single package working on both jvms looks nice. >>> >> >>> >> >>> >> On Sat, Oct 26, 2019 at 4:18 AM Sean Owen <sro...@gmail.com> wrote: >>> >>> >>> >>> I don't think JDK 11 is a separate release (by design). We build >>> >>> everything targeting JDK 8 and it should work on JDK 11 too. >>> >>> >>> >>> So, just two releases, but, frankly I think we soon need to stop >>> >>> multiple releases for multiple Hadoop versions, and stick to Hadoop >>> 3. >>> >>> I think it's fine to try to release for Hadoop 2 as the support still >>> >>> exists, and because the difference happens to be larger due to the >>> >>> different Hive dependency. >>> >>> >>> >>> On Fri, Oct 25, 2019 at 2:08 PM Xingbo Jiang <jiangxb1...@gmail.com> >>> wrote: >>> >>> > >>> >>> > Hi all, >>> >>> > >>> >>> > I would like to bring out a discussion on how many packages shall >>> be released in 3.0.0-preview, the ones I can think of now: >>> >>> > >>> >>> > * scala 2.12 + hadoop 2.7 >>> >>> > * scala 2.12 + hadoop 3.2 >>> >>> > * scala 2.12 + hadoop 3.2 + JDK 11 >>> >>> > >>> >>> > Do you have other combinations to add to the above list? >>> >>> > >>> >>> > Cheers, >>> >>> > >>> >>> > Xingbo >>> >>> >>> >>> --------------------------------------------------------------------- >>> >>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>> >>> >>> >> >>> >> >>> >> -- >>> >> --- >>> >> Takeshi Yamamuro >>> >>> --------------------------------------------------------------------- >>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>> >>>