Awesome!! will give it a try again.  Thanks!!

- Thanks, via mobile,  excuse brevity.
On Jun 19, 2016 11:32 AM, "Xiao Li" <gatorsm...@gmail.com> wrote:

> Hi, Yash,
>
> It should work.
>
>  val df = spark.range(1, 5)
>    .select('id + 1 as 'p1, 'id + 2 as 'p2, 'id + 3 as 'p3, 'id + 4 as 'p4,
> 'id + 5 as 'p5, 'id as 'b)
>    .selectExpr("p1", "p2", "p3", "p4", "p5", "CAST(b AS STRING) AS
> s").coalesce(1)
>
>  df.write.partitionBy("p1", "p2", "p3", "p4",
> "p5").text(dir.getCanonicalPath)
>  val newDF = spark.read.text(dir.getCanonicalPath)
>  newDF.show()
>
>  df.write.partitionBy("p1", "p2", "p3", "p4", "p5")
>    .mode(SaveMode.Append).text(dir.getCanonicalPath)
>  val newDF2 = spark.read.text(dir.getCanonicalPath)
>  newDF2.show()
>
> I tried it. It works well.
>
> Thanks,
>
> Xiao Li
>
> 2016-06-18 8:57 GMT-07:00 Yash Sharma <yash...@gmail.com>:
>
>> Hi All,
>> I have been using the parquet append mode for write which works just
>> fine. Just wanted to check if the same is supported for plain text format.
>> The below code blows up with error saying the file already exists.
>>
>>
>>
>> {code}
>> userEventsDF.write.mode("append").partitionBy("year", "month",
>> "date").text(outputDir)
>> or,
>> userEventsDF.write.mode("append").partitionBy("year", "month",
>> "date").format("text").save(outputDir)
>> {code}
>>
>
>

Reply via email to