Hello!

I want to read csv files with pyspark using (spark_session).read.csv().
There is a whole bunch of nice options, especially an option "locale", nut 
nonetheless a decimal comma instead of a decimal point is not understood when 
reading float/double input even when the locale is set to 'de-DE'. I am using 
spark 3.2.0.
Of course I can read the column as string and write my own float-reader, but 
this will be inefficient in python.
And a simple csv generated by Excel will have decimal commas if written in 
Germany (with German localized Excel).

Markus

Reply via email to