Gerald G Probst created ZEPPELIN-5120:
-----------------------------------------
Summary: Unable to open Spark Interpreter after upgrading from
0.8.1 to 0.9.0-preview2
Key: ZEPPELIN-5120
URL: https://issues.apache.org/jira/browse/ZEPPELIN-5120
Project: Zeppelin
Issue Type: Bug
Components: Interpreters, spark
Affects Versions: 0.9.0
Environment: zeppelin-env.sh:
{code:java}
export MASTER="yarn"
export SPARK_HOME="/usr/lib/spark"
export SPARK_SUBMIT_OPTIONS="--driver-memory 12G --executor-memory 8G"
export AWS_ACCESS_KEY_ID=<key>
export AWS_SECRET_ACCESS_KEY=<secret>
export HADOOP_HOME="/usr/lib/hadoop"
export USE_HADOOP=true
export HADOOP_CONF_DIR="/etc/hadoop/conf"{code}
spark: 2.4.0
scala: 2.11.12
java: 1.8.01_232
OS: Debian 4.9.144-3.1
Reporter: Gerald G Probst
We were successfully running Zeppelin 0.8.1 on a Google Cloud Dataproc cluster
and we needed to upgrade as we were encounter in a bug that was resolved in
newer versions. After upgrading to 0.9.0-preview2 we can no longer connect to
the Spark Interpreter.
Any code returns the following :
{noformat}
org.apache.zeppelin.interpreter.InterpreterException:
org.apache.zeppelin.interpreter.InterpreterException: Fail to open
SparkInterpreter at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:76)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:760)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at
org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
at
org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) Caused by:
org.apache.zeppelin.interpreter.InterpreterException: Fail to open
SparkInterpreter at
org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:122) at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
... 8 more Caused by: java.lang.reflect.InvocationTargetException at
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498) at
org.apache.zeppelin.spark.BaseSparkScalaInterpreter.spark2CreateContext(BaseSparkScalaInterpreter.scala:292)
at
org.apache.zeppelin.spark.BaseSparkScalaInterpreter.createSparkContext(BaseSparkScalaInterpreter.scala:230)
at
org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:98)
at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:106)
... 9 more Caused by: java.lang.VerifyError: Bad return type Exception Details:
Location:
org/apache/hadoop/hdfs/DFSClient.getQuotaUsage(Ljava/lang/String;)Lorg/apache/hadoop/fs/QuotaUsage;
@157: areturn Reason: Type 'org/apache/hadoop/fs/ContentSummary' (current
frame, stack[0]) is not assignable to 'org/apache/hadoop/fs/QuotaUsage' (from
method signature) Current Frame: bci: @157 flags: { } locals: {
'org/apache/hadoop/hdfs/DFSClient', 'java/lang/String',
'org/apache/hadoop/ipc/RemoteException', 'java/io/IOException' } stack: {
'org/apache/hadoop/fs/ContentSummary' } Bytecode: 0x0000000: 2ab6 00df 2a13
01ff 2bb6 00b4 4d01 4e2a 0x0000010: b400 422b b902 0002 003a 042c c600 1d2d
0x0000020: c600 152c b600 b6a7 0012 3a05 2d19 05b6 0x0000030: 00b8 a700 072c
b600 b619 04b0 3a04 1904 0x0000040: 4e19 04bf 3a06 2cc6 001d 2dc6 0015 2cb6
0x0000050: 00b6 a700 123a 072d 1907 b600 b8a7 0007 0x0000060: 2cb6 00b6 1906
bf4d 2c07 bd00 d159 0312 0x0000070: d353 5904 12dd 5359 0512 de53 5906 1302
0x0000080: 0153 b600 d44e 2dc1 0201 9900 14b2 0023 0x0000090: 1302 02b9 002b
0200 2a2b b602 03b0 2dbf 0x00000a0: Exception Handler Table: bci [35, 39] =>
handler: 42 bci [15, 27] => handler: 60 bci [15, 27] => handler: 68 bci [78,
82] => handler: 85 bci [60, 70] => handler: 68 bci [4, 57] => handler: 103 bci
[60, 103] => handler: 103 Stackmap Table:
full_frame(@42,{Object[#736],Object[#759],Object[#814],Object[#784],Object[#1217]},{Object[#784]})
same_frame(@53) same_frame(@57)
full_frame(@60,{Object[#736],Object[#759],Object[#814],Object[#784]},{Object[#784]})
same_locals_1_stack_item_frame(@68,Object[#784])
full_frame(@85,{Object[#736],Object[#759],Object[#814],Object[#784],Top,Top,Object[#784]},{Object[#784]})
same_frame(@96) same_frame(@100)
full_frame(@103,{Object[#736],Object[#759]},{Object[#839]})
append_frame(@158,Object[#839],Object[#799]) at
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:164)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2669) at
org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:94) at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2703) at
org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2685) at
org.apache.hadoop.fs.FileSystem.get(FileSystem.java:373) at
org.apache.spark.util.Utils$.getHadoopFileSystem(Utils.scala:1866) at
org.apache.spark.scheduler.EventLoggingListener.<init>(EventLoggingListener.scala:71)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:528) at
org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2530) at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:935)
at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:926)
at scala.Option.getOrElse(Option.scala:121) at
org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:926)
... 17 more{noformat}
We can connect to spark via the spark-shell and run commands successfully.
This is blocking our current development and have been unable to find a way to
resolve this issue.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)