Dear all,

We are using Zeppelin. And I have added the export
PYTHONPATH=/home/clash/sparks/spark-1.6.1-bin-hadoop12/python
to zeppelin-env.sh.
But each time, when I want to use pyspark, for example the program:

%pyspark
from pyspark import SparkContext
logFile = "hiv.data"
logData = sc.textFile(logFile).cache()
numAs = logData.filter(lambda s: 'a' in s).count()
numBs = logData.filter(lambda s: 'b' in s).count()
print "Lines with a: %i, lines with b: %i" % (numAs, numBs)

It can firstly run well. But second time, I run it again I will get such
error:
*Traceback (most recent call last):*
*  File "/tmp/zeppelin_pyspark-4018989172273347075.py", line 238, in
<module>*
*    sc.setJobGroup(jobGroup, "Zeppelin")*
*  File
"/home/clash/sparks/spark-1.6.1-bin-hadoop12/python/pyspark/context.py",
line 876, in setJobGroup*
*    self._jsc.setJobGroup(groupId, description, interruptOnCancel)*
*AttributeError: 'NoneType' object has no attribute 'setJobGroup'*

*I need to rm */tmp/zeppelin_pyspark-4018989172273347075.py and start
zeppelin again to let it work.
Does anyone have idea why?

Thanks

Reply via email to