Thank you Sam for answer, I have solved problem by loading all decimals columns as string and replacing all commas with dots but this solution is lacking of automatic infer schema which is quite nice functionality.
I can work on adding new option to DataFrameReader for localization like: spark.read.option("NumberLocale", "German").csv("filefromeurope.csv") Just wonder if it will be accepted? Best Regards, Arkadiusz Bicz On Thu, Feb 23, 2017 at 1:35 PM, Sam Elamin <hussam.ela...@gmail.com> wrote: > Hi Arkadiuz > > Not sure if there is a localisation ability but I'm sure other will > correct me if I'm wrong > > > What you could do is write a udf function that replaces the commas with a . > > Assuming you know the column in question > > > Regards > Sam > On Thu, 23 Feb 2017 at 12:31, Arkadiusz Bicz <arkadiusz.b...@gmail.com> > wrote: > >> Hi Team, >> >> I would like to know if it is possible to specify decimal localization >> for DataFrameReader for csv? >> >> I have cvs files from localization where decimal separator is comma like >> 0,32 instead of US way like 0.32 >> >> Is it a way to specify in current version of spark to provide >> localization: >> >> spark.read.option("sep",";").option("header", >> "true").option("inferSchema", "true").format("csv").load(" >> nonuslocalized.csv") >> >> If not should I create ticket in jira for this ? I can work on solution >> if not available. >> >> Best Regards, >> >> Arkadiusz Bicz >> >> >>