[
https://issues.apache.org/jira/browse/SPARK-37198?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17438789#comment-17438789
]
Chuck Connell commented on SPARK-37198:
---------------------------------------
There are many hints/techtips on the Internet which say that
{{file://local_path }}already works to read and write local files from a Spark
cluster. But in my testing (from Databricks) this is not true. I have never
gotten it to work.
If there is already a way to read/write local files, please say the exact,
tested method to do so.
> pyspark.pandas read_csv() and to_csv() should handle local files
> -----------------------------------------------------------------
>
> Key: SPARK-37198
> URL: https://issues.apache.org/jira/browse/SPARK-37198
> Project: Spark
> Issue Type: Sub-task
> Components: PySpark
> Affects Versions: 3.2.0
> Reporter: Chuck Connell
> Priority: Major
>
> Pandas programmers who move their code to Spark would like to import and
> export text files to and from their local disk. I know there are technical
> hurdles to this (since Spark is usually in a cluster that does not know where
> your local computer is) but it would really help code migration.
> For read_csv() and to_csv(), the syntax {{*file://c:/Temp/my_file.csv* }}(or
> something like this) should import and export to the local disk on Windows.
> Similarly for Mac and Linux.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]