Just being too lazy. should define it as custom UDF
def ChangeDate(word : String) : String = {
return
word.substring(6,10)+"-"+word.substring(3,5)+"-"+word.substring(0,2)
}
Register it as custom UDF
sqlContext.udf.register("ChangeDate", ChangeDate(_:String))
And use it in mapping
scala> df
Hi,
I have the following CSV load
val df =
sqlContext.read.format("com.databricks.spark.csv").option("inferSchema",
"true").option("header", "true").load("/data/stg/table2")
I have defined this UDF
def ChangeDate(word : String) : String = {
return
word.substring(6,10)+"-"+word.substring(3,5)