Thank you for your response. Yes, I tried this solution, and it works fine, but this solution for collocated hive cluster. I need to query more then one remote clusters in one spark session, and due to this I need to use connection over jdbc. Maybe you know how to query more then one remote server using your solution?
2017-06-08 20:37 GMT+03:00 Richard Moorhead <richard.moorh...@c2fo.com>: > You might try pointing your spark context at the hive metastore via: > > val conf = new SparkConf() > > conf.set("hive.metastore.uris", "your.thrift.server:9083") > > > val sparkSession = SparkSession.builder() > .config(conf) > .enableHiveSupport() > .getOrCreate() > > > > > . . . . . . . . . . . . . . . . . . . . . . . . . . . > > Richard Moorhead > Software Engineer > richard.moorh...@c2fo.com <richard.moorh...@gmail.com> > > C2FO: The World's Market for Working Capital® > > > <https://www.linkedin.com/company/c2fo?trk=vsrp_companies_res_name&trkInfo=VSRPsearchId%3A125658601427902817660%2CVSRPtargetId%3A1555109%2CVSRPcmpt%3Aprimary> > <https://www.youtube.com/c/C2FOMarket> <https://twitter.com/C2FO> > <https://plus.google.com/+C2foMarket/posts> > <https://www.facebook.com/C2FOMarketplace> > <https://c2fo.com/media-coverage/c2fo-included-forbes-fintech-50> > > The information contained in this message and any attachment may be > privileged, confidential, and protected from disclosure. If you are not the > intended recipient, or an employee, or agent responsible for delivering > this message to the intended recipient, you are hereby notified that any > dissemination, distribution, or copying of this communication is strictly > prohibited. If you have received this communication in error, please notify > us immediately by replying to the message and deleting from your computer. > > > > ------------------------------ > *From:* Даша Ковальчук <dashakovalchu...@gmail.com> > *Sent:* Thursday, June 8, 2017 12:30 PM > *To:* ayan guha > *Cc:* user@spark.apache.org > *Subject:* Re: [Spark Core] Does spark support read from remote Hive > server via JDBC > > The result is count = 0. > > 2017-06-08 19:42 GMT+03:00 ayan guha <guha.a...@gmail.com>: > >> What is the result of test.count()? >> >> On Fri, 9 Jun 2017 at 1:41 am, Даша Ковальчук <dashakovalchu...@gmail.com> >> wrote: >> >>> Thanks for your reply! >>> Yes, I tried this solution and had the same result. Maybe you have >>> another solution or maybe I can execute query in another way on remote >>> cluster? >>> >>> 2017-06-08 18:30 GMT+03:00 Даша Ковальчук <dashakovalchu...@gmail.com>: >>> >>>> Thanks for your reply! >>>> Yes, I tried this solution and had the same result. Maybe you have >>>> another solution or maybe I can execute query in another way on remote >>>> cluster? >>>> >>> >>>> 2017-06-08 18:10 GMT+03:00 Vadim Semenov <vadim.seme...@datadoghq.com>: >>>> >>>>> Have you tried running a query? something like: >>>>> >>>>> ``` >>>>> test.select("*").limit(10).show() >>>>> ``` >>>>> >>>>> On Thu, Jun 8, 2017 at 4:16 AM, Даша Ковальчук < >>>>> dashakovalchu...@gmail.com> wrote: >>>>> >>>>>> Hi guys, >>>>>> >>>>>> I need to execute hive queries on remote hive server from spark, but >>>>>> for some reasons i receive only column names(without data). >>>>>> Data available in table, I checked it via HUE and java jdbc >>>>>> connection. >>>>>> >>>>>> Here is my code example: >>>>>> val test = spark.read >>>>>> .option("url", "jdbc:hive2://remote.hive.server: >>>>>> 10000/work_base") >>>>>> .option("user", "user") >>>>>> .option("password", "password") >>>>>> .option("dbtable", "some_table_with_data") >>>>>> .option("driver", "org.apache.hive.jdbc.HiveDriver") >>>>>> .format("jdbc") >>>>>> .load() >>>>>> test.show() >>>>>> >>>>>> >>>>>> Scala version: 2.11 >>>>>> Spark version: 2.1.0, i also tried 2.1.1 >>>>>> Hive version: CDH 5.7 Hive 1.1.1 >>>>>> Hive JDBC version: 1.1.1 >>>>>> >>>>>> But this problem available on Hive with later versions, too. >>>>>> I didn't find anything in mail group answers and StackOverflow. >>>>>> Could you, please, help me with this issue or could you help me find >>>>>> correct >>>>>> solution how to query remote hive from spark? >>>>>> >>>>>> Thanks in advance! >>>>>> >>>>> >>>>> >>>> -- >> Best Regards, >> Ayan Guha >> > >