By default Spark 1.3 has bindings to Hive 0.13.1 though you can bind it to
Hive 0.12 if you specify it in the profile when building Spark as per
https://spark.apache.org/docs/1.3.0/building-spark.html.
If you are downloading a pre built version of Spark 1.3 - then by default,
it is set to Hive 0.1
Most likely you have an existing Hive installation with data in it. In this
case i was not able to get Spark 1.3 communicate with existing Hive meta
store. Hence when i read any table created in hive, Spark SQL used to
complain "Data table not found"
If you get it working, please share the steps.
Hi,
I use Hive 0.12 for Spark 1.2 at the moment and plan to upgrade to Spark
1.3.x
Could anyone advise which Hive version should be used to match Spark 1.3.x?
Can I use Hive 1.1.0 for Spark 1.3? or can I use Hive 0.14 for Spark 1.3?
Regards
Arthur