Hi Timur, did you use hiveContext or sqlContext or the spark way mentioned in the http://spark.apache.org/docs/latest/sql-programming-guide.html?
Regards, Gourav Sengupta On Wed, Nov 30, 2016 at 5:35 PM, Yin Huai <yh...@databricks.com> wrote: > Hello Michael, > > Thank you for reporting this issue. It will be fixed by > https://github.com/apache/spark/pull/16080. > > Thanks, > > Yin > > On Tue, Nov 29, 2016 at 11:34 PM, Timur Shenkao <t...@timshenkao.su> wrote: > >> Hi! >> >> Do you have real HIVE installation? >> Have you built Spark 2.1 & Spark 2.0 with HIVE support ( -Phive >> -Phive-thriftserver ) ? >> >> It seems that you use "default" Spark's HIVE 1.2.1. Your metadata is >> stored in local Derby DB which is visible to concrete Spark installation >> but not for all. >> >> On Wed, Nov 30, 2016 at 4:51 AM, Michael Allman <mich...@videoamp.com> >> wrote: >> >>> This is not an issue with all tables created in Spark 2.1, though I'm >>> not sure why some work and some do not. I have found that a table created >>> as such >>> >>> sql("create table test stored as parquet as select 1") >>> >>> in Spark 2.1 cannot be read in previous versions of Spark. >>> >>> Michael >>> >>> >>> > On Nov 29, 2016, at 5:15 PM, Michael Allman <mich...@videoamp.com> >>> wrote: >>> > >>> > Hello, >>> > >>> > When I try to read from a Hive table created by Spark 2.1 in Spark 2.0 >>> or earlier, I get an error: >>> > >>> > java.lang.ClassNotFoundException: Failed to load class for data >>> source: hive. >>> > >>> > Is there a way to get previous versions of Spark to read tables >>> written with Spark 2.1? >>> > >>> > Cheers, >>> > >>> > Michael >>> >>> >>> --------------------------------------------------------------------- >>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>> >>> >> >