[ 
https://issues.apache.org/jira/browse/LIVY-859?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyorgy Gal updated LIVY-859:
----------------------------
    Fix Version/s: 0.10.0
                       (was: 0.9.0)

This issue has been moved to the 0.10.0 release as part of a bulk update. If 
you feel this is moved out inappropriately, feel free to provide justification 
and reset the Fix Version to 0.9.0.

> Start PySpark application Failed via sparkmagic with Python 3.7.6 
> ------------------------------------------------------------------
>
>                 Key: LIVY-859
>                 URL: https://issues.apache.org/jira/browse/LIVY-859
>             Project: Livy
>          Issue Type: Question
>         Environment: CDH: 6.3.2
> Spark: 2.4.0
> Python: 3.7.6
> Livy: 0.7.0-incubating
>            Reporter: Eric Luo
>            Priority: Major
>             Fix For: 0.10.0
>
>
> I have two CDH clusters (stage and prod) with the same environment, the same 
> Livy service installed on them, a Python 3.7 environment configured via 
> Ansible, and I write PySpark by calling Livy via sparkmagic in jupyter-lab. 
> code in jupyter-lab, the stage environment works fine, but the prod 
> environment gives an error.
> Error log:
>  
> {code:java}
> 21/05/20 14:43:45 INFO driver.SparkEntries: Created Spark session (with Hive 
> support).
> 21/05/20 14:43:50 ERROR repl.PythonInterpreter: Process has died with 134
> 21/05/20 14:43:50 ERROR repl.PythonInterpreter: Fatal Python error: 
> initfsencoding: unable to load the file system codec
> ModuleNotFoundError: No module named 'encodings'{code}
>  
> On a prod environment machine, I actually have PYSPARK_PYTHON 
> (/opt/miniconda/bin/python) configured to ``import encodings``, and I can run 
> pyspark SHELL directly without any problems.
> My environment configuration.
> CDH: 6.3.2
> Spark: 2.4.0
> Python: 3.7.6
> In CDH's "Spark Service Advanced Configuration Snippet (Safety Valve) for 
> spark-conf/spark-env.sh" configuration, I configured PYSPARK_PYTHON and 
> PYSPARK_ DRIVER_PYTHON
> ```
> export PYSPARK_PYTHON=${PYSPARK_PYTHON:-/opt/miniconda/bin/python}
> export 
> PYSPARK_DRIVER_PYTHON=${PYSPARK_DRIVER_PYTHON:-/opt/miniconda/bin/python}
> ```
> livy-env.sh also configures PYSPARK_PYTHON and PYSPARK_DRIVER_PYTHON
> ```
> PYSPARK_PYTHON=/opt/miniconda/bin/python
> PYSPARK_DRIVER_PYTHON=/opt/miniconda/bin/python
> JAVA_HOME=/usr/java/jdk1.8.0_181-cloudera/
> HADOOP_HOME=/opt/cloudera/parcels/CDH/lib/hadoop
> HADOOP_CONF_DIR=/etc/hadoop/conf
> SPARK_HOME=/opt/cloudera/parcels/CDH/lib/spark
> ```



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to