Hi Mich,

I never managed to run Hive on Spark with a spark master other than local so I 
am afraid I don’t have a reply here.
But do try some things. Firstly, run hive as 

hive --hiveconf hive.root.logger=DEBUG,console

so that you are able to see what the exact error is.

I am afraid I cannot be much of a help as I think I reached the same point 
(where it would work only when setting spark.master=local) before abandoning.

Cheers



> On 27 Nov 2015, at 01:59, Mich Talebzadeh <m...@peridale.co.uk> wrote:
> 
> Hi Sophia,
>  
>  
> There is no Hadoop-2.6. I believe you should use Hadoop-2.4 as shown below
>  
>  
> mvn -Phadoop-2.4 -Dhadoop.version=2.6.0 -DskipTests clean package
>  
> Also if you are building it for Hive on Spark engine, you should not include 
> Hadoop.jar files in your build.
>  
> For example I tried to build spark 1.3 from source code (I read that this 
> version works OK with Hive, having tried unsuccessfully spark 1.5.2). 
>  
> The following command created the tar file
>  
> ./make-distribution.sh --name "hadoop2-without-hive" --tgz 
> "-Pyarn,hadoop-provided,hadoop-2.4,parquet-provided"
>  
> spark-1.3.0-bin-hadoop2-without-hive.tar.gz
>  
>  
> Now I have other issues making Hive to use Spark execution engine (requires 
> Hive 1.1 or above )
>  
> In hive I do
>  
> set spark.home=/usr/lib/spark;
> set hive.execution.engine=spark;
> set spark.master=spark://127.0.0.1:7077 <spark://127.0.0.1:7077>;
> set spark.eventLog.enabled=true;
> set spark.eventLog.dir=/usr/lib/spark/logs;
> set spark.executor.memory=512m;
> set spark.serializer=org.apache.spark.serializer.KryoSerializer;
> use asehadoop;
> select count(1) from t;
>  
> I get the following
>  
> OK
> Time taken: 0.753 seconds
> Query ID = hduser_20151127003523_e9863e84-9a81-4351-939c-36b3bef36478
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=<number>
> Failed to execute spark task, with exception 
> 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark 
> client.)'
> FAILED: Execution Error, return code 1 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask
>  
> HTH,
>  
> Mich
>  
> NOTE: The information in this email is proprietary and confidential. This 
> message is for the designated recipient only, if you are not the intended 
> recipient, you should destroy it immediately. Any information in this message 
> shall not be understood as given or endorsed by Peridale Technology Ltd, its 
> subsidiaries or their employees, unless expressly so stated. It is the 
> responsibility of the recipient to ensure that this email is virus free, 
> therefore neither Peridale Ltd, its subsidiaries nor their employees accept 
> any responsibility.
>  
> From: Sofia [mailto:sofia.panagiot...@taiger.com] 
> Sent: 18 November 2015 16:50
> To: user@hive.apache.org
> Subject: Hive version with Spark
>  
> Hello
>  
> After various failed tries to use my Hive (1.2.1) with my Spark (Spark 1.4.1 
> built for Hadoop 2.2.0) I decided to try to build again Spark with Hive.
> I would like to know what is the latest Hive version that can be used to 
> build Spark at this point.
>  
> When downloading Spark 1.5 source and trying:
>  
> mvn -Pyarn -Phadoop-2.6 -Dhadoop.version=2.6.0 -Phive -Phive-1.2.1 
> -Phive-thriftserver  -DskipTests clean package
>  
> I get :
>  
> The requested profile "hive-1.2.1" could not be activated because it does not 
> exist.
>  
> Thank you
> Sofia

Reply via email to