Hi Sofia,
Thank you for the detailed guide. 
I confirmed that there are compatibility issues between Hive 1.2.1 and Spark 
1.5.X. Hive 1.2.1 is by default built against spark 1.3.1 and compatible with 
spark 1.4.1,but doesn't work with 1.5.X.


Basically,I am using somewhat the same configuration as yours,but I am using 
yarn-client as the spark.master.
Did you try to use yarn-client as the spark.master?


I am building spark1.3.1 without hive and hive-thriftserver. I will get back to 
you when I do the test,


Thank you, Sofia. 





At 2016-01-11 23:56:32, "Sofia" <sofia.panagiot...@taiger.com> wrote:

From what I have seen you need to use a Spark version that does not contain the 
Hive jars. (maybe build one as described at the page).
Maybe someone can respond to the compatibility issue better than me. 


However, this is how I managed to make hive work with Spark


I had similar issues with the downloaded Spark 1.5.2 and in the end I used the 
prebuilt Spark 1.4.1 with my Hive 1.2.1 (and Hadoop 2.7.1).
When doing the setup I used the configurations below:


hive-env.sh


export SPARK_HOME=/usr/local/spark
export HADOOP_HOME=/usr/local/hadoop


hive-site.xml


<property>
   <name>hive.execution.engine</name>
   <value>spark</value>
 </property>
 <property>
   <name>spark.master</name>
   <value>spark://spark-master:7077</value>
 </property>




spark-env.sh


export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin
export SPARK_HOME=/usr/local/spark


export SPARK_DIST_CLASSPATH=`hadoop classpath`
export SPARK_CLASSPATH=$CLASSPATH




I don’t know the answer to your issue exactly as at some point I faced the 
same, but I hope this might help.
Cheers




On 11 Jan 2016, at 11:39, Todd <bit1...@163.com> wrote:




Thank you, Sofia.


From the log, looks it is java.lang.AbstractMethodError that leads to the job 
failure.
I am using Hive 1.2.1+Spark 1.5.2, is this a compatibility  issue?








16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildPlan 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Map 1 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO Configuration.deprecation: mapred.task.is.map is deprecated. 
Instead, use mapreduce.task.ismap
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Processing alias t1
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Adding input file 
hdfs://hadoop.bit.com:9000/user/hive/warehouse/db_1.db/t1
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Serializing MapWork via kryo
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1452505326307 end=1452505326332 duration=25 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: ensureFreeSpace(505800) called with 
curMem=0, maxMem=278302556
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory 
(estimated size 493.9 KB, free 264.9 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: ensureFreeSpace(42143) called with 
curMem=505800, maxMem=278302556
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 41.2 KB, free 264.9 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10.12.167.42:38883 (size: 41.2 KB, free: 265.4 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 ERROR util.Utils: uncaught error in thread SparkListenerBus, stopping 
SparkContext
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.AbstractMethodError
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:62)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:56)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:79)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1136)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO spark.SparkContext: Created broadcast 0 from hadoopRDD at 
SparkPlanGenerator.java:188
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/metrics/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/api,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/static,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Map 1 
start=1452505326274 end=1452505326824 duration=550 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Reducer 2 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Serializing ReduceWork via kryo
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO ui.SparkUI: Stopped Spark web UI at http://10.12.167.42:4040
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO scheduler.DAGScheduler: Stopping DAGScheduler
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Interrupting monitor thread
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Asking each executor to shut 
down
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Stopped
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1452505326849 end=1452505326874 duration=25 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Reducer 2 
start=1452505326824 end=1452505326936 duration=112 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkBuildPlan 
start=1452505326273 end=1452505326936 duration=663 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildRDDGraph 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:07 INFO log.PerfLogger: </PERFLOG method=SparkBuildRDDGraph 
start=1452505326936 end=1452505327000 duration=64 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:07 INFO client.RemoteDriver: Failed to run job 
b5a8886e-8229-4773-b943-b3c6f3c710bf
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.SparkContext.submitJob(SparkContext.scala:1981)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:118)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:116)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDD.withScope(RDD.scala:310)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions.foreachAsync(AsyncRDDActions.scala:116)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.api.java.JavaRDDLike$class.foreachAsync(JavaRDDLike.scala:690)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.api.java.AbstractJavaRDDLike.foreachAsync(JavaRDDLike.scala:47)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient$JobStatusJob.call(RemoteHiveSparkClient.java:257)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:366)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:335)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.FutureTask.run(FutureTask.java:262)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.lang.Thread.run(Thread.java:745)
16/01/11 17:42:07 [RPC-Handler-3]: INFO client.SparkClientImpl: Received result 
for b5a8886e-8229-4773-b943-b3c6f3c710bf
Status: Failed
16/01/11 17:42:07 [main]: ERROR status.SparkJobMonitor: Status: Failed





At 2016-01-11 17:17:51, "Sofia" <sofia.panagiot...@taiger.com> wrote:

Check log files to see the exact error. Alternatively run hive like this, to 
get the exact error


hive --hiveconf hive.root.logger=INFO,console 


There is a good chance that you are encountering problems between the Hive and 
Spark versions and installation.
See 
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started




On 11 Jan 2016, at 08:47, Todd <bit1...@163.com> wrote:


Hive,


I am trying out the Hive on Spark with hive 1.2.1 and spark 1.5.2. Could 
someone help me on this? Thanks!


Following are my steps:
1. build spark 1.5.2 without Hive and Hive Thrift Server. At this point,  I can 
use it to submit application using spark-submit --master yarn-client
2. And the built spark assembly jar into $HIVE_HOME/lib
3. start hive and add the following parameters
hive > set spark.master=yarn-client
hive > set spark.executor.memory=512M
hive > set spark.driver.memory=512M
hive > set  spark.executor.instances=1


4. Then I run a simple query : select count(1) from t1;


The job fails will following error:


===============================================================================
YARN executor launch context:
  env:
    CLASSPATH -> 
{{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
    SPARK_LOG_URL_STDERR -> 
http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stderr?start=-4096
    SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1452320323183_0007
    SPARK_YARN_CACHE_FILES_FILE_SIZES -> 142746538
    SPARK_USER -> hadoop
    SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE
    SPARK_YARN_MODE -> true
    SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1452496343550
    SPARK_LOG_URL_STDOUT -> 
http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stdout?start=-4096
    SPARK_YARN_CACHE_FILES -> 
hdfs://hadoop.bit.com:9000/user/hadoop/.sparkStaging/application_1452320323183_0007/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar


  command:
    {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms512m 
-Xmx512m 
'-Dhive.spark.log.dir=/home/hadoop/software/bigdata/spark-1.5.2-bin-hadoop2.6.0-withouthive/logs/'
 -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.driver.port=43675' 
-Dspark.yarn.app.container.log.dir=<LOG_DIR> 
org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url 
akka.tcp://sparkDriver@10.12.167.42:43675/user/CoarseGrainedScheduler 
--executor-id 2 --hostname hadoop-Aspire-TC-606 --cores 1 --app-id 
application_1452320323183_0007 --user-class-path file:$PWD/__app__.jar 1> 
<LOG_DIR>/stdout 2> <LOG_DIR>/stderr
===============================================================================
      
16/01/11 15:12:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy : 
hadoop-Aspire-TC-606:50804
16/01/11 15:12:40 INFO yarn.YarnAllocator: Completed container 
container_1452320323183_0007_01_000003 (state: COMPLETE, exit status: 1)
16/01/11 15:12:40 INFO yarn.YarnAllocator: Container marked as failed: 
container_1452320323183_0007_01_000003. Exit status: 1. Diagnostics: Exception 
from container-launch.
Container id: container_1452320323183_0007_01_000003
Exit code: 1
Stack trace: ExitCodeException exitCode=1: 
at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
at org.apache.hadoop.util.Shell.run(Shell.java:455)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:211)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)




Container exited with a non-zero exit code 1



































Reply via email to