I was able to install version 0.5.0 and it works great with JupyterHub!
Thanks again for the help! However, when I tried to install the PySpark
and the SparkR kernel (using the command line option
--interpreters=Scala,PySpark,SparkR,SQL), I got the following error
messages:
[ToreeInstall] ERROR | Unknown interpreter PySpark. Skipping
installation of PySpark interpreter
[ToreeInstall] ERROR | Unknown interpreter SparkR. Skipping installation
of SparkR interpreter
Are there two kernels available in 0.5.0?
On 2/17/22 8:30 PM, Kevin Bates wrote:
Hello,
You should checkout v0.5.0-rc5:
https://github.com/apache/incubator-toree/releases/tag/v0.5.0-incubating-rc5
which includes support for Spark 3.2.
On 2022/02/17 22:27:30 Artemis User wrote:
After looking at the Toree's Readme on github, I realized that the toree
version 0.4.x only supports Spark version 2.x, whereas the master branch
supports Spark 3.2.x. Could someone confirm this? In addition, is a
distribution package of toree from the master branch available?
Thanks!
On 2/17/22 4:57 PM, Artemis User wrote:
Hi Toree Dev Team, Could someone please help with resolving the
following error when starting the Toree Scala Kernel inside Jupyter?
My configuration settings:
* JupyterHub version 3.2.9
* OpenJDK 11
* Spark 3.2.0 with Scala version 2.12
* Apache Toree 0.4.0-incubating
Thanks a lot for your help!
Exception in thread "main" java.lang.NoClassDefFoundError:
scala/App$class
at org.apache.toree.Main$.<init>(Main.scala:24)
at org.apache.toree.Main$.<clinit>(Main.scala)
at org.apache.toree.Main.main(Main.scala)
at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
Method)
at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
at
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955)
at
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
at
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.ClassNotFoundException: scala.App$class
at
java.base/java.net.URLClassLoader.findClass(URLClassLoader.java:476)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:589)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522)
... 15 more