Moon, This is what I have in my configuration
export ZEPPELIN_INTERPRETERS=org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.hive.HiveInterpreter export ZEPPELIN_INTERPRETER_DIR=/home/sambit/incubator-zeppelin/interpreter export ZEPPELIN_PORT=8901 export HADOOP_CONF_DIR=/usr/lib/hadoop/etc/hadoop export SPARK_YARN_JAR=/usr/lib/spark/lib/spark-assembly-1.2.0-cdh5.3.0-hadoop2.5.0-cdh5.3.0.jar export ZEPPELIN_NOTEBOOK_DIR=/home/sambit/zep-notebook-dir # Where notebook saved Used this command mvn install -DskipTests -Dspark.version=1.2.1 -Dhadoop.version=2.5.0 to build Zeppelin as provided in the website That’s all. Should the –Dhadoop.version change to 2.5.0-cdh5.3.0? Regards, Sambit. From: moon soo Lee [mailto:[email protected]] Sent: Thursday, April 30, 2015 5:25 PM To: [email protected] Subject: Re: Scheduler already terminated error Hi, That error message can be shown when Zeppelin fails to create SparkContext. Could you check Zeppelin configuration for your yarn cluster? How did you setup Zeppelin for your Yarn cluster? Like Zeppelin build command against your spark / hadoop version, Zeppelin Interpreter setting, hadoop/yarn configuration files. Thanks, moon On Fri, May 1, 2015 at 8:02 AM Sambit Tripathy (RBEI/EDS1) <[email protected]<mailto:[email protected]>> wrote: Hi, After installation, I tried to run this simple spark command and got this error. Any idea what it could be? Command: %spark val ctx = new org.apache.spark.sql.SqlContext(sc) Error: Scheduler already terminated org.apache.zeppelin.scheduler.RemoteScheduler.submit(RemoteScheduler.java:122) org.apache.zeppelin.notebook.Note.run(Note.java:271) org.apache.zeppelin.socket.NotebookServer.runParagraph(NotebookServer.java:531) org.apache.zeppelin.socket.NotebookServer.onMessage(NotebookServer.java:119) org.java_websocket.server.WebSocketServer.onWebsocketMessage(WebSocketServer.java:469) org.java_websocket.WebSocketImpl.decodeFrames(WebSocketImpl.java:368) org.java_websocket.WebSocketImpl.decode(WebSocketImpl.java:157) org.java_websocket.server.WebSocketServer$WebSocketWorker.run(WebSocketServer.java:657) ERROR What is the best way to verify that Spark Interpreter is working correctly? Is this a Yarn error? PS: I am using yarn. Regards, Sambit.
