Fixed this by setting fileformat -> "parquet" On Mon, Oct 22, 2018 at 11:48 AM Nikhil Goyal <nownik...@gmail.com> wrote:
> Hi guys, > > My code is failing with this error > > java.lang.Exception: S2V: FATAL ERROR for job S2V_job9197956021769393773. > Job status information is available in the Vertica table > test.S2V_JOB_STATUS_USER_NGOYAL. Unable to save intermediate orc files to > HDFS path:hdfs://hadoop-dw2-nn.smf1.com/tmp/S2V_job9197956021769393773. > *Error message:org.apache.spark.sql.AnalysisException: The ORC data source > must be used with Hive support enabled;* > > This is how I am writing it. I followed steps from this link > <https://www.vertica.com/blog/integrating-apache-spark/>. > > dataFrame > .write > .format("com.vertica.spark.datasource.DefaultSource") > .options(connectionProperties) > .mode(SaveMode.Append) > .save() > > Does anybody have any idea how to fix this? > > > Thanks > Nikhil > >