Did you export SPARK_HOME in conf/zeppelin-env.sh? Could you verify the some code works with ${SPARK_HOME}/bin/pyspark, on the same machine that zeppelin runs?
Thanks, moon On Wed, Sep 14, 2016 at 8:07 AM Abhi Basu <9000r...@gmail.com> wrote: > Oops sorry. the above code generated this error: > > RROR [2016-09-14 10:04:27,121] ({qtp2003293121-11} > NotebookServer.java[onMessage]:221) - Can't handle message > org.apache.zeppelin.interpreter.InterpreterException: > org.apache.thrift.transport.TTransportException > at > org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel(RemoteInterpreter.java:319) > at > org.apache.zeppelin.interpreter.LazyOpenInterpreter.cancel(LazyOpenInterpreter.java:100) > at org.apache.zeppelin.notebook.Paragraph.jobAbort(Paragraph.java:330) > at org.apache.zeppelin.scheduler.Job.abort(Job.java:239) > at > org.apache.zeppelin.socket.NotebookServer.cancelParagraph(NotebookServer.java:995) > at > org.apache.zeppelin.socket.NotebookServer.onMessage(NotebookServer.java:180) > at > org.apache.zeppelin.socket.NotebookSocket.onWebSocketText(NotebookSocket.java:56) > at > org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextMessage(JettyListenerEventDriver.java:128) > at > org.eclipse.jetty.websocket.common.message.SimpleTextMessage.messageComplete(SimpleTextMessage.java:69) > at > org.eclipse.jetty.websocket.common.events.AbstractEventDriver.appendMessage(AbstractEventDriver.java:65) > at > org.eclipse.jetty.websocket.common.events.JettyListenerEventDriver.onTextFrame(JettyListenerEventDriver.java:122) > at > org.eclipse.jetty.websocket.common.events.AbstractEventDriver.incomingFrame(AbstractEventDriver.java:161) > at > org.eclipse.jetty.websocket.common.WebSocketSession.incomingFrame(WebSocketSession.java:309) > at > org.eclipse.jetty.websocket.common.extensions.ExtensionStack.incomingFrame(ExtensionStack.java:214) > at org.eclipse.jetty.websocket.common.Parser.notifyFrame(Parser.java:220) > at org.eclipse.jetty.websocket.common.Parser.parse(Parser.java:258) > at > org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection.readParse(AbstractWebSocketConnection.java:632) > at > org.eclipse.jetty.websocket.common.io.AbstractWebSocketConnection.onFillable(AbstractWebSocketConnection.java:480) > at > org.eclipse.jetty.io.AbstractConnection$2.run(AbstractConnection.java:544) > at > org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:635) > at > org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:555) > at java.lang.Thread.run(Thread.java:745) > Caused by: org.apache.thrift.transport.TTransportException > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > at > org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429) > at > org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318) > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219) > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69) > at > org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.recv_cancel(RemoteInterpreterService.java:274) > at > org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.cancel(RemoteInterpreterService.java:259) > at > org.apache.zeppelin.interpreter.remote.RemoteInterpreter.cancel(RemoteInterpreter.java:316) > ... 21 more > > > This is my spark interpreter settings: > > > spark %spark , %spark.pyspark , %spark.r , %spark.sql , %spark.dep > Option > Interpreter for note > > Connect to existing process > Properties > name value > args > master yarn-client > spark.app.name Zeppelin > spark.cores.max > spark.executor.memory > zeppelin.R.cmd R > zeppelin.R.image.width 100% > zeppelin.R.knitr true > zeppelin.R.render.options out.format = 'html', comment = NA, echo = > FALSE, results = 'asis', message = F, warning = F > zeppelin.dep.additionalRemoteRepository spark-packages, > http://dl.bintray.com/spark-packages/maven,false; > zeppelin.dep.localrepo local-repo > zeppelin.interpreter.localRepo > /usr/local/bin/zeppelin-0.6.1-bin-all/local-repo/2BXF675WU > zeppelin.pyspark.python python > zeppelin.spark.concurrentSQL false > zeppelin.spark.importImplicit true > zeppelin.spark.maxResult 1000 > zeppelin.spark.printREPLOutput true > zeppelin.spark.sql.stacktrace false > zeppelin.spark.useHiveContext true > > > On Wed, Sep 14, 2016 at 10:05 AM, Abhi Basu <9000r...@gmail.com> wrote: > >> %pyspark >> >> input_file = "hdfs:////tmp/filenname.gz" >> >> raw_rdd = sc.textFile(input_file) >> >> >> > > > -- > Abhi Basu >