Hi, I am new to using Spark and Parquet files,
Below is what i am trying to do, on Spark-shell, val df = sqlContext.parquetFile("/data/LM/Parquet/Segment/pages/part-m-00000.gz.parquet") Have also tried below command, val df=sqlContext.read.format("parquet").load("/data/LM/Parquet/Segment/pages/part-m-00000.gz.parquet") Now i have an other existing parquet file to which i want to append this Parquet file data of df. so i use, df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet","parquet", SaveMode.Append ) also tried below command, df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet", SaveMode.Append ) and it throws me below error, <console>:26: error: not found: value SaveMode df.save("/data/LM/Parquet/Segment/pages2/part-m-00000.gz.parquet","parquet", SaveMode.Append ) Please help me, in case i am doing something wrong here. Regards, Satyajit.