Supreeth Sharma created ZEPPELIN-3433:
-----------------------------------------
Summary: Incorrect status shown for %spark2 interpreter
Key: ZEPPELIN-3433
URL: https://issues.apache.org/jira/browse/ZEPPELIN-3433
Project: Zeppelin
Issue Type: Bug
Components: zeppelin-server
Affects Versions: 0.8.0
Reporter: Supreeth Sharma
Fix For: 0.8.0
%spark2 interpreter is showing the status as 'Finished' even when it has
actually failed.
Steps to reproduce :
1) Create spark.conf with incorrect jar file.
{code:java}
%spark2.conf
spark.jars /tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar
spark.app.name test_new_1
spark.executor.instances 2
{code}
2) Now create spark context by running any spark statement.
{code:java}
%spark
sc.version
{code}
3) Though spark context is not created and para didnt run successfully, the
status is shown as 'Finished' without showing error message
Below exceptions are seen in zeppelin logs, but are not shown on the UI.
{code:java}
INFO [2018-04-27 14:58:55,014] ({pool-3-thread-2} Logging.scala[logInfo]:54) -
Bound SparkUI to 0.0.0.0, and started at
http://ctr-e138-1518143905142-267605-01-000002.hwx.site:4040
ERROR [2018-04-27 14:58:55,030] ({pool-3-thread-2} Logging.scala[logError]:91)
- Failed to add /tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar to
Spark environment
java.io.FileNotFoundException: Jar
/tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar not found
at org.apache.spark.SparkContext.addJarFile$1(SparkContext.scala:1814)
at org.apache.spark.SparkContext.addJar(SparkContext.scala:1842)
at
org.apache.spark.SparkContext$$anonfun$12.apply(SparkContext.scala:457)
at
org.apache.spark.SparkContext$$anonfun$12.apply(SparkContext.scala:457)
at scala.collection.immutable.List.foreach(List.scala:381)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:457)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493)
at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930)
at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921)
at scala.Option.getOrElse(Option.scala:121)
at
org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:44)
at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:39)
at
org.apache.zeppelin.spark.OldSparkInterpreter.createSparkSession(OldSparkInterpreter.java:345)
at
org.apache.zeppelin.spark.OldSparkInterpreter.getSparkSession(OldSparkInterpreter.java:219)
at
org.apache.zeppelin.spark.OldSparkInterpreter.open(OldSparkInterpreter.java:738)
at
org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:61)
at
org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
at
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:611)
at org.apache.zeppelin.scheduler.Job.run(Job.java:186)
at
org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
WARN [2018-04-27 14:58:55,074] ({pool-3-thread-2}
Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so
jobs will be scheduled in FIFO order. To use fair scheduling, configure pools
in fairscheduler.xml or set spark.scheduler.allocation.file to a file that
contains the configuration.
{code}
Correct status(ERROR) is shown for livy2 interpreter in this scenario. Issue is
happening only with spark2 interpreter.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)