Nassir created ZEPPELIN-2678:
--------------------------------
Summary: Pyspark cell fails to execute, but normal spark code in
scala executes fine
Key: ZEPPELIN-2678
URL: https://issues.apache.org/jira/browse/ZEPPELIN-2678
Project: Zeppelin
Issue Type: Bug
Reporter: Nassir
Hi,
I have installed zeppelin for windows and can now run cells with the default
spark interpreter - so scala code.
However, when I try to execute a pyspark cell e.g.
%pyspark
x = 5
I get an error:
"failed to start pyspark"
Any ideas on what is going wrong here? I can see %pyspark as an interpreter
under Spark in the Interpreter page.
Do I need to set some environment variables? I had anaconda installed for
running python but did not add any environmental path variables?
Thanks
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)