Can you check your PATH environment variable in both the console and in Jupyter? From Python, you can do that like this:
import os os.environ['PATH'] On 11 May 2017 at 13:24, David Arenburg <[email protected]> wrote: > Hello all, > > I've been trying to figure this out for a week with no success. > > I'm simply trying to initiate SparkContext in Jupyter but getting the > following error when running SparkConf() : > > > ``` > from pyspark import SparkConf > SparkConf() > > ``` > > ---------------------------------------------------------------------------OSError > Traceback (most recent call > last)<ipython-input-12-0c80a6a098f7> in <module>() 1 #import statements > 2 from pyspark import SparkConf----> 3 SparkConf() 4 > /root/david/spark/python/pyspark/conf.pyc in __init__(self, loadDefaults, > _jvm, _jconf) 102 else: 103 from pyspark.context > import SparkContext--> 104 SparkContext._ensure_initialized() > 105 _jvm = _jvm or SparkContext._jvm 106 > self._jconf = _jvm.SparkConf(loadDefaults) > /root/david/spark/python/pyspark/context.pyc in _ensure_initialized(cls, > instance, gateway) 241 with SparkContext._lock: 242 > if not SparkContext._gateway:--> 243 SparkContext._gateway = > gateway or launch_gateway() 244 SparkContext._jvm = > SparkContext._gateway.jvm 245 > /root/david/spark/python/pyspark/java_gateway.pyc in launch_gateway() 74 > def preexec_func(): 75 > signal.signal(signal.SIGINT, signal.SIG_IGN)---> 76 proc = > Popen(command, stdin=PIPE, preexec_fn=preexec_func, env=env) 77 > else: 78 # preexec_fn not supported on Windows > /mnt/anaconda/lib/python2.7/subprocess.pyc in __init__(self, args, bufsize, > executable, stdin, stdout, stderr, preexec_fn, close_fds, shell, cwd, env, > universal_newlines, startupinfo, creationflags) 709 > p2cread, p2cwrite, 710 c2pread, > c2pwrite,--> 711 errread, errwrite) 712 > except Exception: 713 # Preserve original exception in > case os.close raises. > /mnt/anaconda/lib/python2.7/subprocess.pyc in _execute_child(self, args, > executable, preexec_fn, close_fds, cwd, env, universal_newlines, startupinfo, > creationflags, shell, to_close, p2cread, p2cwrite, c2pread, c2pwrite, > errread, errwrite) 1341 raise 1342 > child_exception = pickle.loads(data)-> 1343 raise > child_exception 1344 1345 > OSError: [Errno 2] No such file or directory > > > > This is my startJupyter.sh script which I use to lunch Jupyter > > #!/bin/bash > if ps -ef |grep $USER| grep python > /dev/null > then > echo "Jupyter is Running - Restarting" > echo "Killing jupyter-notebook process" > > running_id=$(ps -ef |grep $USER| grep python) > stringarray=($running_id) > echo ${stringarray[1]} > kill -9 ${stringarray[1]} > > export SPARK_HOME='/usr/lib/spark/' > export PYTHONPATH=$SPARK_HOME/python/:$PYTHONPATH:$SPARK_HOME/ > python/lib/py4j-0.9-src.zip > > #jupyter nbextension enable --py widgetsnbextension > /mnt/anaconda/bin/jupyter notebook & > > else > echo "Jupyter is Not Running" > echo "Starting Jupyter-NoteBook" > export SPARK_HOME='/usr/lib/spark/' > export PYTHONPATH=$SPARK_HOME/python/:$PYTHONPATH:$SPARK_HOME/ > python/lib/py4j-0.9-src.zip > > #jupyter nbextension enable --py widgetsnbextension > /mnt/anaconda/bin/jupyter notebook & > fi > > > When running the same in python in console (not in Jupyter), it works fine > > Python 2.7.12 |Anaconda 4.2.0 (64-bit)| (default, Jul 2 2016, 17:42:40) > [GCC 4.4.7 20120313 (Red Hat 4.4.7-1)] on linux2 > Type "help", "copyright", "credits" or "license" for more information. > Anaconda is brought to you by Continuum Analytics. > Please check out: http://continuum.io/thanks and https://anaconda.org > >>> from pyspark import SparkConf > >>> SparkConf() > Using Spark's default log4j profile: org/apache/spark/log4j- > defaults.properties > Setting default log level to "WARN". > To adjust logging level use sc.setLogLevel(newLevel). > <pyspark.conf.SparkConf object at 0x7f482f78b6d0> > > > I've validated both the python version and the module path in console and > in Jupyter and they seem to match > > >>> import sys > >>> sys.version > '2.7.12 |Anaconda 4.2.0 (64-bit)| (default, Jul 2 2016, 17:42:40) \n[GCC > 4.4.7 20120313 (Red Hat 4.4.7-1)]' > > >>> import inspect > >>> import pyspark > >>> inspect.getfile(pyspark) > '/root/david/spark/python/pyspark/__init__.pyc' > > I can't think of anything else I could do wrong, please help > Thanks, David > > My specs: > > NAME="Amazon Linux AMI" > VERSION="2017.03" > ID="amzn" > ID_LIKE="rhel fedora" > VERSION_ID="2017.03" > PRETTY_NAME="Amazon Linux AMI 2017.03" > ANSI_COLOR="0;33" > CPE_NAME="cpe:/o:amazon:linux:2017.03:ga" > HOME_URL="http://aws.amazon.com/amazon-linux-ami/" > Amazon Linux AMI release 2017.03 > > -- > You received this message because you are subscribed to the Google Groups > "Project Jupyter" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To post to this group, send email to [email protected]. > To view this discussion on the web visit https://groups.google.com/d/ > msgid/jupyter/a1924dad-a893-447f-8b79-735bb2ff0fb8%40googlegroups.com > <https://groups.google.com/d/msgid/jupyter/a1924dad-a893-447f-8b79-735bb2ff0fb8%40googlegroups.com?utm_medium=email&utm_source=footer> > . > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "Project Jupyter" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/jupyter/CAOvn4qhABomX5qJWnG%2BJzCe-Rx%3D2wUrOswuRmucpD49td0drkg%40mail.gmail.com. For more options, visit https://groups.google.com/d/optout.
