Hi Timur,

did you use hiveContext or sqlContext or the spark way mentioned in the
http://spark.apache.org/docs/latest/sql-programming-guide.html?


Regards,
Gourav Sengupta

On Wed, Nov 30, 2016 at 5:35 PM, Yin Huai <yh...@databricks.com> wrote:

> Hello Michael,
>
> Thank you for reporting this issue. It will be fixed by
> https://github.com/apache/spark/pull/16080.
>
> Thanks,
>
> Yin
>
> On Tue, Nov 29, 2016 at 11:34 PM, Timur Shenkao <t...@timshenkao.su> wrote:
>
>> Hi!
>>
>> Do you have real HIVE installation?
>> Have you built Spark 2.1 & Spark 2.0 with HIVE support ( -Phive
>> -Phive-thriftserver ) ?
>>
>> It seems that you use "default" Spark's HIVE 1.2.1. Your metadata is
>> stored in local Derby DB which is visible to concrete Spark installation
>> but not for all.
>>
>> On Wed, Nov 30, 2016 at 4:51 AM, Michael Allman <mich...@videoamp.com>
>> wrote:
>>
>>> This is not an issue with all tables created in Spark 2.1, though I'm
>>> not sure why some work and some do not. I have found that a table created
>>> as such
>>>
>>> sql("create table test stored as parquet as select 1")
>>>
>>> in Spark 2.1 cannot be read in previous versions of Spark.
>>>
>>> Michael
>>>
>>>
>>> > On Nov 29, 2016, at 5:15 PM, Michael Allman <mich...@videoamp.com>
>>> wrote:
>>> >
>>> > Hello,
>>> >
>>> > When I try to read from a Hive table created by Spark 2.1 in Spark 2.0
>>> or earlier, I get an error:
>>> >
>>> > java.lang.ClassNotFoundException: Failed to load class for data
>>> source: hive.
>>> >
>>> > Is there a way to get previous versions of Spark to read tables
>>> written with Spark 2.1?
>>> >
>>> > Cheers,
>>> >
>>> > Michael
>>>
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>
>>>
>>
>

Reply via email to