Hello, Ranadip!
I tried your solution, but still have no results. Also I didn’t find
anything in logs.
Kerberos disabled, dfs.permissions = false.
Thanks.

2017-06-08 20:52 GMT+03:00 Ranadip Chatterjee <ranadi...@gmail.com>:

> Looks like your session user does not have the required privileges on the
> remote hdfs directory that is holding the hive data. Since you get the
> columns, your session is able to read the metadata, so connection to the
> remote hiveserver2 is successful. You should be able to find more
> troubleshooting information in the remote hiveserver2 log file.
>
> Try with select* limit 10 to keep it simple.
>
> Ranadip
>
>
> On 8 Jun 2017 6:31 pm, "Даша Ковальчук" <dashakovalchu...@gmail.com>
> wrote:
>
> The  result is count = 0.
>
> 2017-06-08 19:42 GMT+03:00 ayan guha <guha.a...@gmail.com>:
>
>> What is the result of test.count()?
>>
>> On Fri, 9 Jun 2017 at 1:41 am, Даша Ковальчук <dashakovalchu...@gmail.com>
>> wrote:
>>
>>> Thanks for your reply!
>>> Yes, I tried this solution and had the same result. Maybe you have
>>> another solution or maybe I can execute query in another way on remote
>>> cluster?
>>>
>>> 2017-06-08 18:30 GMT+03:00 Даша Ковальчук <dashakovalchu...@gmail.com>:
>>>
>>>> Thanks for your reply!
>>>> Yes, I tried this solution and had the same result. Maybe you have
>>>> another solution or maybe I can execute query in another way on remote
>>>> cluster?
>>>>
>>>
>>>> 2017-06-08 18:10 GMT+03:00 Vadim Semenov <vadim.seme...@datadoghq.com>:
>>>>
>>>>> Have you tried running a query? something like:
>>>>>
>>>>> ```
>>>>> test.select("*").limit(10).show()
>>>>> ```
>>>>>
>>>>> On Thu, Jun 8, 2017 at 4:16 AM, Даша Ковальчук <
>>>>> dashakovalchu...@gmail.com> wrote:
>>>>>
>>>>>> Hi guys,
>>>>>>
>>>>>> I need to execute hive queries on remote hive server from spark, but
>>>>>> for some reasons i receive only column names(without data).
>>>>>> Data available in table, I checked it via HUE and java jdbc
>>>>>>  connection.
>>>>>>
>>>>>> Here is my code example:
>>>>>> val test = spark.read
>>>>>>         .option("url", "jdbc:hive2://remote.hive.server:
>>>>>> 10000/work_base")
>>>>>>         .option("user", "user")
>>>>>>         .option("password", "password")
>>>>>>         .option("dbtable", "some_table_with_data")
>>>>>>         .option("driver", "org.apache.hive.jdbc.HiveDriver")
>>>>>>         .format("jdbc")
>>>>>>         .load()
>>>>>> test.show()
>>>>>>
>>>>>>
>>>>>> Scala version: 2.11
>>>>>> Spark version: 2.1.0, i also tried 2.1.1
>>>>>> Hive version: CDH 5.7 Hive 1.1.1
>>>>>> Hive JDBC version: 1.1.1
>>>>>>
>>>>>> But this problem available on Hive with later versions, too.
>>>>>> I didn't find anything in mail group answers and StackOverflow.
>>>>>> Could you, please, help me with this issue or could you help me find 
>>>>>> correct
>>>>>> solution how to query remote hive from spark?
>>>>>>
>>>>>> Thanks in advance!
>>>>>>
>>>>>
>>>>>
>>>> --
>> Best Regards,
>> Ayan Guha
>>
>
>
>

Reply via email to