Re: Support for decimal separator (comma or period) in spark 2.1

2017-02-23 Thread Hyukjin Kwon
Please take a look at https://issues.apache.org/jira/browse/SPARK-18359. 2017-02-23 21:53 GMT+09:00 Arkadiusz Bicz : > Thank you Sam for answer, I have solved problem by loading all decimals > columns as string and replacing all commas with dots but this solution is > lacking of automatic infer s

Re: Support for decimal separator (comma or period) in spark 2.1

2017-02-23 Thread Arkadiusz Bicz
Thank you Sam for answer, I have solved problem by loading all decimals columns as string and replacing all commas with dots but this solution is lacking of automatic infer schema which is quite nice functionality. I can work on adding new option to DataFrameReader for localization like: spark.r

Re: Support for decimal separator (comma or period) in spark 2.1

2017-02-23 Thread Sam Elamin
Hi Arkadiuz Not sure if there is a localisation ability but I'm sure other will correct me if I'm wrong What you could do is write a udf function that replaces the commas with a . Assuming you know the column in question Regards Sam On Thu, 23 Feb 2017 at 12:31, Arkadiusz Bicz wrote: > Hi T

Re: Support for decimal separator (comma or period) in spark 2.1

2017-02-23 Thread Arkadiusz Bicz
> > Hi Team, > > I would like to know if it is possible to specify decimal localization for > DataFrameReader for csv? > > I have cvs files from localization where decimal separator is comma like > 0,32 instead of US way like 0.32 > > Is it a way to specify in current version of spark to provide >