Hello,

Just to be clear, I am not explicitly building the Hadoop edition. I get the 
described error when I simply specify the Spark version:

  mvn clean package -Dspark.version=2.0.0 -DskipTests

Thanks for the quick response!


-----Original Message-----
From: Sergey Kozlov [mailto:skoz...@gridgain.com] 
Sent: 19 September 2016 05:56
To: dev@ignite.apache.org
Subject: Re: Support for Spark 2.0

Hi

It's a known issue IGNITE-3596 Hadoop edition can't be compiled against spark 
2.0.0 <https://issues.apache.org/jira/browse/IGNITE-3596>

Unfortunately there's no progress yet

On Mon, Sep 19, 2016 at 1:36 PM, Phadnis, Varun <phad...@sky.optymyze.com>
wrote:

> Hello,
>
> Can someone please tell me when is the support for Spark 2.0 planned?
>
> Currently Ignite cannot be built for Spark 2.0. Attempting this yields 
> the following error :
>
>
> ----------------------------------------------------------------------
> --
>
> [INFO] Building ignite-spark 1.8.0-SNAPSHOT [INFO]
>
> ----------------------------------------------------------------------
> --
>
> [INFO]
>
> [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ ignite-spark 
> --- [INFO] Deleting /home/spark/code/ignite/modules/spark/target
>
> [INFO]
>
> [INFO] --- flatten-maven-plugin:1.0.0-beta-3:clean 
> (flatten.clean.before) @ ignite-spark --- [INFO] Deleting 
> /home/spark/code/ignite/ modules/spark/pom-installed.xml
>
> [INFO]
>
> [INFO] --- maven-enforcer-plugin:1.4:enforce (default) @ ignite-spark 
> --- [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process 
> (default) @ ignite-spark --- [INFO] [INFO] --- 
> maven-remote-resources-plugin:1.5:process
> (ignite-dependencies) @ ignite-spark --- [INFO] [INFO] --- 
> maven-resources-plugin:2.6:resources (default-resources) @ 
> ignite-spark
> --- [INFO] Using 'UTF-8' encoding to copy filtered resources.
>
> [INFO] skip non existing resourceDirectory /home/spark/code/ignite/ 
> modules/spark/src/main/resources
>
> [INFO] Copying 4 resources
>
> [INFO] Copying 4 resources
>
> [INFO]
>
> [INFO] --- flatten-maven-plugin:1.0.0-beta-3:flatten (flatten) @ 
> ignite-spark --- [INFO] Generating flattened POM of project 
> org.apache.ignite:ignite-spark:jar:1.8.0-SNAPSHOT...
>
> [WARNING] Ignoring multiple XML header comment!
>
> [INFO]
>
> [INFO] --- scala-maven-plugin:3.2.0:add-source (scala-compile-first) @ 
> ignite-spark --- [INFO] Add Source directory:
>
> /home/spark/code/ignite/modules/spark/src/main/scala
>
> [INFO] Add Test Source directory:
>
> /home/spark/code/ignite/modules/spark/src/test/scala
>
> [INFO]
>
> [INFO] --- scala-maven-plugin:3.2.0:compile (scala-compile-first) @ 
> ignite-spark --- [WARNING]  Expected all dependencies to require Scala
> version: 2.11.7 [WARNING]  
> org.apache.ignite:ignite-spark:1.8.0-SNAPSHOT
> requires scala
>
> version: 2.11.7
>
> [WARNING]  com.twitter:chill_2.11:0.8.0 requires scala version: 2.11.7 
> [WARNING]  org.apache.spark:spark-core_2.11:2.0.0 requires scala version:
>
> 2.11.8
>
> [WARNING] Multiple versions of scala libraries detected!
>
> [INFO] /home/spark/code/ignite/modules/spark/src/main/scala:-1: info:
>
> compiling
>
> [INFO] Compiling 8 source files to
>
> /home/spark/code/ignite/modules/spark/target/classes at 1474028222935 
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:25:
>
> error: object Logging is not a member of package org.apache.spark 
> [ERROR] import org.apache.spark.{Logging, SparkContext}
>
> [ERROR]        ^
>
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:37:
>
> error: not found: type Logging
>
> [ERROR]     ) extends Serializable with Logging {
>
> [ERROR]                                 ^
>
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:50:
>
> error: not found: value logInfo
>
> [ERROR]         logInfo("Will start Ignite nodes on " + workers + "
>
> workers")
>
> [ERROR]         ^
>
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:129:
>
> error: not found: value logInfo
>
> [ERROR]             logInfo("Setting IGNITE_HOME from driver not as it is
>
> not available on this worker: " + igniteHome)
>
> [ERROR]             ^
>
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:146:
>
> error: not found: value logError
>
> [ERROR]                 logError("Failed to start Ignite.", e)
>
> [INFO]                 ^
>
> [ERROR]
>
> /home/spark/code/ignite/modules/spark/src/main/scala/
> org/apache/ignite/spark/IgniteContext.scala:164:
>
> error: not found: value logInfo
>
> [ERROR]                 logInfo("Will stop Ignite nodes on " + workers + "
>
> workers")
>
> [ERROR]                 ^
>
> [ERROR] 6 errors found
>
> [INFO]
>
> ----------------------------------------------------------------------
> --
>
> Thanks!
>



--
Sergey Kozlov
GridGain Systems
www.gridgain.com

Reply via email to