As to the spark versions that are supported. Spark has made non-compatible API changes in 1.5, and that's the reason why Hive 1.1.0 doesn't work with Spark 1.5. However, the latest Hive in master or branch-1 should work with spark 1.5.
Also, later CDH 5.4.x versions have already supported Spark 1.5. CDH 5.7, which is coming so, will support Spark 1.6. --Xuefu On Tue, Dec 15, 2015 at 3:50 PM, Mich Talebzadeh <m...@peridale.co.uk> wrote: > To answer your point: > > > > “why would spark 1.5.2 specifically would not work with hive?” > > > > Because I tried Spark 1.5.2 and it did not work and unfortunately the only > version seem to work (albeit requires messaging around) is version 1.3.1 of > Spark. > > > > Look at the threads on “Managed to make Hive run on Spark engine” in > user@hive.apache.org > > > > > > HTH, > > > > > > Mich Talebzadeh > > > > *Sybase ASE 15 Gold Medal Award 2008* > > A Winning Strategy: Running the most Critical Financial Data on ASE 15 > > > http://login.sybase.com/files/Product_Overviews/ASE-Winning-Strategy-091908.pdf > > Author of the books* "A Practitioner’s Guide to Upgrading to Sybase ASE > 15", ISBN 978-0-9563693-0-7*. > > co-author *"Sybase Transact SQL Guidelines Best Practices", ISBN > 978-0-9759693-0-4* > > *Publications due shortly:* > > *Complex Event Processing in Heterogeneous Environments*, ISBN: > 978-0-9563693-3-8 > > *Oracle and Sybase, Concepts and Contrasts*, ISBN: 978-0-9563693-1-4, volume > one out shortly > > > > http://talebzadehmich.wordpress.com > > > > NOTE: The information in this email is proprietary and confidential. This > message is for the designated recipient only, if you are not the intended > recipient, you should destroy it immediately. Any information in this > message shall not be understood as given or endorsed by Peridale Technology > Ltd, its subsidiaries or their employees, unless expressly so stated. It is > the responsibility of the recipient to ensure that this email is virus > free, therefore neither Peridale Ltd, its subsidiaries nor their employees > accept any responsibility. > > > > *From:* Ophir Etzion [mailto:op...@foursquare.com] > *Sent:* 15 December 2015 22:42 > *To:* user@hive.apache.org > *Cc:* u...@spark.apache.org > *Subject:* Re: Hive on Spark - Error: Child process exited before > connecting back > > > > Hi, > > the versions are spark 1.3.0 and hive 1.1.0 as part of cloudera 5.4.3. > > I find it weird that it would work only on the version you mentioned as > there is documentation (not good documentation but still..) on how to do it > with cloudera that packages different versions. > > Thanks for the answer though. > > why would spark 1.5.2 specifically would not work with hive? > > > > Ophir > > > > On Tue, Dec 15, 2015 at 5:33 PM, Mich Talebzadeh <m...@peridale.co.uk> > wrote: > > Hi, > > > > The only version that I have managed to run Hive using Spark engine is > Spark 1.3.1 on Hive 1.2.1 > > > > Can you confirm the version of Spark you are running? > > > > FYI, Spark 1.5.2 will not work with Hive. > > > > HTH > > > > Mich Talebzadeh > > > > *Sybase ASE 15 Gold Medal Award 2008* > > A Winning Strategy: Running the most Critical Financial Data on ASE 15 > > > http://login.sybase.com/files/Product_Overviews/ASE-Winning-Strategy-091908.pdf > > Author of the books* "A Practitioner’s Guide to Upgrading to Sybase ASE > 15", ISBN 978-0-9563693-0-7*. > > co-author *"Sybase Transact SQL Guidelines Best Practices", ISBN > 978-0-9759693-0-4* > > *Publications due shortly:* > > *Complex Event Processing in Heterogeneous Environments*, ISBN: > 978-0-9563693-3-8 > > *Oracle and Sybase, Concepts and Contrasts*, ISBN: 978-0-9563693-1-4, volume > one out shortly > > > > http://talebzadehmich.wordpress.com > > > > NOTE: The information in this email is proprietary and confidential. This > message is for the designated recipient only, if you are not the intended > recipient, you should destroy it immediately. Any information in this > message shall not be understood as given or endorsed by Peridale Technology > Ltd, its subsidiaries or their employees, unless expressly so stated. It is > the responsibility of the recipient to ensure that this email is virus > free, therefore neither Peridale Ltd, its subsidiaries nor their employees > accept any responsibility. > > > > *From:* Ophir Etzion [mailto:op...@foursquare.com] > *Sent:* 15 December 2015 22:27 > *To:* u...@spark.apache.org; user@hive.apache.org > *Subject:* Hive on Spark - Error: Child process exited before connecting > back > > > > Hi, > > > > when trying to do Hive on Spark on CDH5.4.3 I get the following error when > trying to run a simple query using spark. > > I've tried setting everything written here ( > https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started) > as well as what the cdh recommends. > > any one encountered this as well? (searching for it didn't help much) > > the error: > > ERROR : Failed to execute spark task, with exception > 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark > client.)' > > org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark > client. > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:57) > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:114) > > at > org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:120) > > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:97) > > at > org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88) > > at > org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1640) > > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1399) > > at > org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1183) > > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1049) > > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1044) > > at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:144) > > at > org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:69) > > at > org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:196) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:415) > > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > > at > org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:208) > > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.RuntimeException: > java.util.concurrent.ExecutionException: java.lang.RuntimeException: Cancel > client '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited > before connecting back > > at > com.google.common.base.Throwables.propagate(Throwables.java:156) > > at > org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:109) > > at > org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFactory.java:80) > > at > org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:91) > > at > org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:65) > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:55) > > ... 22 more > > Caused by: java.util.concurrent.ExecutionException: > java.lang.RuntimeException: Cancel client > '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited before > connecting back > > at > io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37) > > at > org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:99) > > ... 26 more > > Caused by: java.lang.RuntimeException: Cancel client > '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited before > connecting back > > at > org.apache.hive.spark.client.rpc.RpcServer.cancelClient(RpcServer.java:179) > > at > org.apache.hive.spark.client.SparkClientImpl$3.run(SparkClientImpl.java:427) > > ... 1 more > > > > ERROR : Failed to execute spark task, with exception > 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark > client.)' > > org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark > client. > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:57) > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:114) > > at > org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:120) > > at > org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:97) > > at > org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > > at > org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88) > > at > org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1640) > > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1399) > > at > org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1183) > > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1049) > > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1044) > > at > org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:144) > > at > org.apache.hive.service.cli.operation.SQLOperation.access$100(SQLOperation.java:69) > > at > org.apache.hive.service.cli.operation.SQLOperation$1$1.run(SQLOperation.java:196) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:415) > > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) > > at > org.apache.hive.service.cli.operation.SQLOperation$1.run(SQLOperation.java:208) > > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > > at java.lang.Thread.run(Thread.java:745) > > Caused by: java.lang.RuntimeException: > java.util.concurrent.ExecutionException: java.lang.RuntimeException: Cancel > client '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited > before connecting back > > at > com.google.common.base.Throwables.propagate(Throwables.java:156) > > at > org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:109) > > at > org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFactory.java:80) > > at > org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:91) > > at > org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:65) > > at > org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:55) > > ... 22 more > > Caused by: java.util.concurrent.ExecutionException: > java.lang.RuntimeException: Cancel client > '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited before > connecting back > > at > io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37) > > at > org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:99) > > ... 26 more > > Caused by: java.lang.RuntimeException: Cancel client > '2b2d7314-e0cc-4933-82a1-992a3299d109'. Error: Child process exited before > connecting back > > at > org.apache.hive.spark.client.rpc.RpcServer.cancelClient(RpcServer.java:179) > > at > org.apache.hive.spark.client.SparkClientImpl$3.run(SparkClientImpl.java:427) > > ... 1 more > > Error: Error while processing statement: FAILED: Execution Error, return > code 1 from org.apache.hadoop.hive.ql.exec.spark.SparkTask > (state=08S01,code=1) > > > > >