Hello,

Yes, I used hiveContext, sqlContext, sparkSession from Java, Scala, Python.
Via spark-shell, spark-submit, IDE (PyCharm, Intellij IDEA).
Everything is perfect because I have Hadoop cluster with configured & tuned
HIVE.

The reason of Michael's error is usually misconfigured or absent HIVE.
Or may be absence of hive-site.xml in $SPARK_HOME/conf/ directory.

On Wed, Nov 30, 2016 at 9:30 PM, Gourav Sengupta <gourav.sengu...@gmail.com>
wrote:

> Hi Timur,
>
> did you use hiveContext or sqlContext or the spark way mentioned in the
> http://spark.apache.org/docs/latest/sql-programming-guide.html?
>
>
> Regards,
> Gourav Sengupta
>
> On Wed, Nov 30, 2016 at 5:35 PM, Yin Huai <yh...@databricks.com> wrote:
>
>> Hello Michael,
>>
>> Thank you for reporting this issue. It will be fixed by
>> https://github.com/apache/spark/pull/16080.
>>
>> Thanks,
>>
>> Yin
>>
>> On Tue, Nov 29, 2016 at 11:34 PM, Timur Shenkao <t...@timshenkao.su>
>> wrote:
>>
>>> Hi!
>>>
>>> Do you have real HIVE installation?
>>> Have you built Spark 2.1 & Spark 2.0 with HIVE support ( -Phive
>>> -Phive-thriftserver ) ?
>>>
>>> It seems that you use "default" Spark's HIVE 1.2.1. Your metadata is
>>> stored in local Derby DB which is visible to concrete Spark installation
>>> but not for all.
>>>
>>> On Wed, Nov 30, 2016 at 4:51 AM, Michael Allman <mich...@videoamp.com>
>>> wrote:
>>>
>>>> This is not an issue with all tables created in Spark 2.1, though I'm
>>>> not sure why some work and some do not. I have found that a table created
>>>> as such
>>>>
>>>> sql("create table test stored as parquet as select 1")
>>>>
>>>> in Spark 2.1 cannot be read in previous versions of Spark.
>>>>
>>>> Michael
>>>>
>>>>
>>>> > On Nov 29, 2016, at 5:15 PM, Michael Allman <mich...@videoamp.com>
>>>> wrote:
>>>> >
>>>> > Hello,
>>>> >
>>>> > When I try to read from a Hive table created by Spark 2.1 in Spark
>>>> 2.0 or earlier, I get an error:
>>>> >
>>>> > java.lang.ClassNotFoundException: Failed to load class for data
>>>> source: hive.
>>>> >
>>>> > Is there a way to get previous versions of Spark to read tables
>>>> written with Spark 2.1?
>>>> >
>>>> > Cheers,
>>>> >
>>>> > Michael
>>>>
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>
>>>>
>>>
>>
>

Reply via email to