ivanthewebber opened a new issue, #1581:
URL: https://github.com/apache/sedona/issues/1581

   ## Expected behavior
   Instructions in 
[docs](https://sedona.apache.org/1.6.1/setup/install-python/) with latest 
versions should succeed without errors. I have been unable to initialize 
Sedona-Spark for the Python API. I think the docs need updated or there are 
errors in the most recent versions.
   
   I installed Sedona and PySpark (with Hadoop) from PyPi and have a Java 11 
JDK and Scala 2.12 on my computer. I also tried installing Spark from a 
download directly. I have tried manually downloading the Jars as well.
   
   I want to initialize the session like follows:
   ```python
   import pyspark
   import pyspark.version
   import pyspark.sql
   import sedona
   import sedona.spark
   
   def get_sedona_spark(spark_version=pyspark.version.__version__, 
scala_version='2.12', sedona_version=sedona.version, geotools_version='28.2') 
-> sedona.spark.SedonaContext:
       """
       Get the Sedona Spark context.
   
       We use the newest version, so Sedona's methods will expect lon-lat order.
       """
   
       if spark_version.count('.') > 1:
           spark_version = '.'.join(spark_version.split('.')[:2])
   
       builder: pyspark.sql.SparkSession.Builder = 
sedona.spark.SedonaContext.builder()
       spark = builder\
           .config(
               'spark.jars.packages',
               
f'org.apache.sedona:sedona-spark-{spark_version}_{scala_version}:{sedona_version},'
 +
               
f'org.datasyslab:geotools-wrapper:{sedona_version}-{geotools_version}'
           ).config(
               'spark.jars.repositories',
               'https://artifacts.unidata.ucar.edu/repository/unidata-all'
           ).getOrCreate()
   
       return sedona.spark.SedonaContext.create(spark)
   
   if __name__ == "__main__":
       get_sedona_spark()
   ```
   
   Ideally like the quickstarts for Spark/Flink there would be simple steps to 
run a simple word count program.
   
   ## Actual behavior
   Various errors. I've tried a lot of variations and recommended fixes from 
StackOverflows but haven't made much progress.
   
   I get errors like the following: `PySparkRuntimeError: [JAVA_GATEWAY_EXITED] 
Java gateway process exited before sending its port number.`
   
   ## Steps to reproduce the problem
   ```
   # create new env
   python -m env env
   ./env/scripts/Activate.ps1
   python -m pip install --upgrade pip
   pip install --upgrade apache-sedona[spark] pyspark
   
   # I tried setting the spark home to a few different options, but
   # if I'm reading the docs right when installing from PyPi I shouldn't need to
   # $SPARK_HOME = "venv/.../pyspark"
   
   # attempt to initialize session (see above)
   python test.py
   ```
   
   ## Settings
   
   Sedona version = 1.6.1, 1.5
   
   Apache Spark version = 3.5.2, 3.5.1, 3.4
   
   API type = Python
   
   Scala version = 2.12
   
   JRE version = 1.11
   
   Python version = 3.12
   
   Environment = Local
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: dev-unsubscr...@sedona.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org

Reply via email to