You need to import org.apache.spark.sql.SaveMode

Cheng

On 7/31/15 6:26 AM, satyajit vegesna wrote:
Hi,

I am new to using Spark and Parquet files,

Below is what i am trying to do, on Spark-shell,

val df = sqlContext.parquetFile("/data/LM/Parquet/Segment/pages/part-m-00000.gz.parquet")
Have also tried below command,

val df=sqlContext.read.format("parquet").load("/data/LM/Parquet/Segment/pages/part-m-00000.gz.parquet")

Now i have an other existing parquet file to which i want to append this Parquet file data of df.

so i use,

df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet","parquet", SaveMode.Append )

also tried below command,

df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet", SaveMode.Append )


and it throws me below error,

<console>:26: error: not found: value SaveMode
df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet","parquet", SaveMode.Append )

Please help me, in case i am doing something wrong here.

Regards,
Satyajit.





---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to