Thanks!
I got this to work.
val csvRdd = sc.parallelize(data.split("\n"))
val df = new
com.databricks.spark.csv.CsvParser().withUseHeader(true).withInferSchema(true).csvRdd(sqlContext,
csvRdd)
> On Apr 15, 2016, at 1:14 PM, Hyukjin Kwon wrote:
>
> Hi,
>
> Would you try
Is this right?
import com.databricks.spark.csv
val csvRdd = data.flatMap(x => x.split("\n"))
val df = new CsvParser().csvRdd(sqlContext, csvRdd, useHeader = true)
Thanks,
Ben
> On Apr 15, 2016, at 1:14 PM, Hyukjin Kwon wrote:
>
> Hi,
>
> Would you try this codes
Hi,
Would you try this codes below?
val csvRDD = ...your processimg for csv rdd..
val df = new CsvParser().csvRdd(sqlContext, csvRDD, useHeader = true)
Thanks!
On 16 Apr 2016 1:35 a.m., "Benjamin Kim" wrote:
> Hi Hyukjin,
>
> I saw that. I don’t know how to use it. I’m
Hi Hyukjin,
I saw that. I don’t know how to use it. I’m still learning Scala on my own. Can
you help me to start?
Thanks,
Ben
> On Apr 15, 2016, at 8:02 AM, Hyukjin Kwon wrote:
>
> I hope it was not too late :).
>
> It is possible.
>
> Please check csvRdd api here,
>
Does anyone know if this is possible? I have an RDD loaded with rows of CSV
data strings. Each string representing the header row and multiple rows of data
along with delimiters. I would like to feed each thru a CSV parser to convert
the data into a dataframe and, ultimately, UPSERT a