hi. I tested your problem in the same as your environment. I think you don't need to set PYSPARK_PYTHON so can you remove export PYSPARK_PYTHON=/home/jin/spark/python in your zeppelin-env.sh? and retry please.
let me share my configuration fyi. export PYTHONPATH=/usr/bin/python export SPARK_HOME=/usr/lib/spark-2.1.0-bin-hadoop2.7 2017-02-24 20:03 GMT+09:00 Beth Lee <[email protected]>: > Yes, I already registered it. > But the result is same. > > Thanks, > Jin > > 2017-02-24 18:05 GMT+09:00 Hyung Sung Shim <[email protected]>: > >> hello. >> Could you set spark-2.1.0-bin-hadoop2.7 path as SPARK_HOME ? >> You can refer to http://zeppelin.apache.org/ >> docs/0.7.0/interpreter/spark.html#1-export-spark_home. >> >> 2017-02-24 15:44 GMT+09:00 Beth Lee <[email protected]>: >> >>> I installed spark-2.1.0-bin-hadoop2.7.tgz and zeppelin-0.7.0-bin-all.tgz in >>> ubuntu. >>> >>> I set the zeppelin-env.sh like below. >>> >>> export PYTHONPATH=/usr/bin/python >>> export PYSPARK_PYTHON=/home/jin/spark/python >>> >>> So I try to use pyspark in the zeppelin notebook. >>> >>> %spark.pyspark >>> >>> print(2+2) >>> >>> There are errors occurred in zeppelin notebook. >>> >>> java.lang.NullPointerException >>> at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) >>> at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:380) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:369) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:144) >>> at >>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:817) >>> at >>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>> at >>> org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:546) >>> at >>> org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:206) >>> at >>> org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:160) >>> at >>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>> at >>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:482) >>> at org.apache.zeppelin.scheduler.Job.run(Job.java:175) >>> at >>> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) >>> at >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>> at >>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>> at >>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> at java.lang.Thread.run(Thread.java:745) >>> >>> I don't know why these errors are occurred. >>> >>> Would you give me some advice? >>> >> >> >
