Hello, Yes, I used hiveContext, sqlContext, sparkSession from Java, Scala, Python. Via spark-shell, spark-submit, IDE (PyCharm, Intellij IDEA). Everything is perfect because I have Hadoop cluster with configured & tuned HIVE.
The reason of Michael's error is usually misconfigured or absent HIVE. Or may be absence of hive-site.xml in $SPARK_HOME/conf/ directory. On Wed, Nov 30, 2016 at 9:30 PM, Gourav Sengupta <gourav.sengu...@gmail.com> wrote: > Hi Timur, > > did you use hiveContext or sqlContext or the spark way mentioned in the > http://spark.apache.org/docs/latest/sql-programming-guide.html? > > > Regards, > Gourav Sengupta > > On Wed, Nov 30, 2016 at 5:35 PM, Yin Huai <yh...@databricks.com> wrote: > >> Hello Michael, >> >> Thank you for reporting this issue. It will be fixed by >> https://github.com/apache/spark/pull/16080. >> >> Thanks, >> >> Yin >> >> On Tue, Nov 29, 2016 at 11:34 PM, Timur Shenkao <t...@timshenkao.su> >> wrote: >> >>> Hi! >>> >>> Do you have real HIVE installation? >>> Have you built Spark 2.1 & Spark 2.0 with HIVE support ( -Phive >>> -Phive-thriftserver ) ? >>> >>> It seems that you use "default" Spark's HIVE 1.2.1. Your metadata is >>> stored in local Derby DB which is visible to concrete Spark installation >>> but not for all. >>> >>> On Wed, Nov 30, 2016 at 4:51 AM, Michael Allman <mich...@videoamp.com> >>> wrote: >>> >>>> This is not an issue with all tables created in Spark 2.1, though I'm >>>> not sure why some work and some do not. I have found that a table created >>>> as such >>>> >>>> sql("create table test stored as parquet as select 1") >>>> >>>> in Spark 2.1 cannot be read in previous versions of Spark. >>>> >>>> Michael >>>> >>>> >>>> > On Nov 29, 2016, at 5:15 PM, Michael Allman <mich...@videoamp.com> >>>> wrote: >>>> > >>>> > Hello, >>>> > >>>> > When I try to read from a Hive table created by Spark 2.1 in Spark >>>> 2.0 or earlier, I get an error: >>>> > >>>> > java.lang.ClassNotFoundException: Failed to load class for data >>>> source: hive. >>>> > >>>> > Is there a way to get previous versions of Spark to read tables >>>> written with Spark 2.1? >>>> > >>>> > Cheers, >>>> > >>>> > Michael >>>> >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >>>> >>>> >>> >> >