Thank you, Sofia.


From the log, looks it is java.lang.AbstractMethodError that leads to the job 
failure.
I am using Hive 1.2.1+Spark 1.5.2, is this a compatibility  issue?








16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildPlan 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Map 1 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO Configuration.deprecation: mapred.task.is.map is deprecated. 
Instead, use mapreduce.task.ismap
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Processing alias t1
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Adding input file 
hdfs://hadoop.bit.com:9000/user/hive/warehouse/db_1.db/t1
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Serializing MapWork via kryo
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1452505326307 end=1452505326332 duration=25 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: ensureFreeSpace(505800) called with 
curMem=0, maxMem=278302556
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: Block broadcast_0 stored as values in memory 
(estimated size 493.9 KB, free 264.9 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: ensureFreeSpace(42143) called with 
curMem=505800, maxMem=278302556
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.MemoryStore: Block broadcast_0_piece0 stored as bytes in 
memory (estimated size 41.2 KB, free 264.9 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO storage.BlockManagerInfo: Added broadcast_0_piece0 in memory on 
10.12.167.42:38883 (size: 41.2 KB, free: 265.4 MB)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 ERROR util.Utils: uncaught error in thread SparkListenerBus, stopping 
SparkContext
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.AbstractMethodError
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:62)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:56)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:79)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1136)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO spark.SparkContext: Created broadcast 0 from hadoopRDD at 
SparkPlanGenerator.java:188
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/metrics/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/api,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/static,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/threadDump,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/executors,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/environment,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/rdd,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/storage,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/pool,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/stage,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/stages,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/job,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs/json,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO handler.ContextHandler: stopped 
o.s.j.s.ServletContextHandler{/jobs,null}
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Map 1 
start=1452505326274 end=1452505326824 duration=550 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkCreateTran.Reducer 2 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=serializePlan 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO exec.Utilities: Serializing ReduceWork via kryo
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO ui.SparkUI: Stopped Spark web UI at http://10.12.167.42:4040
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO scheduler.DAGScheduler: Stopping DAGScheduler
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Shutting down all executors
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Interrupting monitor thread
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Asking each executor to shut 
down
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO cluster.YarnClientSchedulerBackend: Stopped
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=serializePlan 
start=1452505326849 end=1452505326874 duration=25 
from=org.apache.hadoop.hive.ql.exec.Utilities>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkCreateTran.Reducer 2 
start=1452505326824 end=1452505326936 duration=112 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: </PERFLOG method=SparkBuildPlan 
start=1452505326273 end=1452505326936 duration=663 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlanGenerator>
16/01/11 17:42:06 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:06 INFO log.PerfLogger: <PERFLOG method=SparkBuildRDDGraph 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:07 INFO log.PerfLogger: </PERFLOG method=SparkBuildRDDGraph 
start=1452505326936 end=1452505327000 duration=64 
from=org.apache.hadoop.hive.ql.exec.spark.SparkPlan>
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 16/01/11 
17:42:07 INFO client.RemoteDriver: Failed to run job 
b5a8886e-8229-4773-b943-b3c6f3c710bf
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: 
java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.SparkContext.submitJob(SparkContext.scala:1981)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:118)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions$$anonfun$foreachAsync$1.apply(AsyncRDDActions.scala:116)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.RDD.withScope(RDD.scala:310)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.rdd.AsyncRDDActions.foreachAsync(AsyncRDDActions.scala:116)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.api.java.JavaRDDLike$class.foreachAsync(JavaRDDLike.scala:690)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.spark.api.java.AbstractJavaRDDLike.foreachAsync(JavaRDDLike.scala:47)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient$JobStatusJob.call(RemoteHiveSparkClient.java:257)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:366)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:335)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.FutureTask.run(FutureTask.java:262)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
16/01/11 17:42:07 [stderr-redir-1]: INFO client.SparkClientImpl: at 
java.lang.Thread.run(Thread.java:745)
16/01/11 17:42:07 [RPC-Handler-3]: INFO client.SparkClientImpl: Received result 
for b5a8886e-8229-4773-b943-b3c6f3c710bf
Status: Failed
16/01/11 17:42:07 [main]: ERROR status.SparkJobMonitor: Status: Failed





At 2016-01-11 17:17:51, "Sofia" <sofia.panagiot...@taiger.com> wrote:

Check log files to see the exact error. Alternatively run hive like this, to 
get the exact error


hive --hiveconf hive.root.logger=INFO,console 


There is a good chance that you are encountering problems between the Hive and 
Spark versions and installation.
See 
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started




On 11 Jan 2016, at 08:47, Todd <bit1...@163.com> wrote:


Hive,


I am trying out the Hive on Spark with hive 1.2.1 and spark 1.5.2. Could 
someone help me on this? Thanks!


Following are my steps:
1. build spark 1.5.2 without Hive and Hive Thrift Server. At this point,  I can 
use it to submit application using spark-submit --master yarn-client
2. And the built spark assembly jar into $HIVE_HOME/lib
3. start hive and add the following parameters
hive > set spark.master=yarn-client
hive > set spark.executor.memory=512M
hive > set spark.driver.memory=512M
hive > set  spark.executor.instances=1


4. Then I run a simple query : select count(1) from t1;


The job fails will following error:


===============================================================================
YARN executor launch context:
  env:
    CLASSPATH -> 
{{PWD}}<CPS>{{PWD}}/__spark__.jar<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/*
    SPARK_LOG_URL_STDERR -> 
http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stderr?start=-4096
    SPARK_YARN_STAGING_DIR -> .sparkStaging/application_1452320323183_0007
    SPARK_YARN_CACHE_FILES_FILE_SIZES -> 142746538
    SPARK_USER -> hadoop
    SPARK_YARN_CACHE_FILES_VISIBILITIES -> PRIVATE
    SPARK_YARN_MODE -> true
    SPARK_YARN_CACHE_FILES_TIME_STAMPS -> 1452496343550
    SPARK_LOG_URL_STDOUT -> 
http://hadoop-Aspire-TC-606:8042/node/containerlogs/container_1452320323183_0007_01_000003/hadoop/stdout?start=-4096
    SPARK_YARN_CACHE_FILES -> 
hdfs://hadoop.bit.com:9000/user/hadoop/.sparkStaging/application_1452320323183_0007/spark-assembly-1.5.2-hadoop2.6.0.jar#__spark__.jar


  command:
    {{JAVA_HOME}}/bin/java -server -XX:OnOutOfMemoryError='kill %p' -Xms512m 
-Xmx512m 
'-Dhive.spark.log.dir=/home/hadoop/software/bigdata/spark-1.5.2-bin-hadoop2.6.0-withouthive/logs/'
 -Djava.io.tmpdir={{PWD}}/tmp '-Dspark.driver.port=43675' 
-Dspark.yarn.app.container.log.dir=<LOG_DIR> 
org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url 
akka.tcp://sparkDriver@10.12.167.42:43675/user/CoarseGrainedScheduler 
--executor-id 2 --hostname hadoop-Aspire-TC-606 --cores 1 --app-id 
application_1452320323183_0007 --user-class-path file:$PWD/__app__.jar 1> 
<LOG_DIR>/stdout 2> <LOG_DIR>/stderr
===============================================================================
      
16/01/11 15:12:37 INFO impl.ContainerManagementProtocolProxy: Opening proxy : 
hadoop-Aspire-TC-606:50804
16/01/11 15:12:40 INFO yarn.YarnAllocator: Completed container 
container_1452320323183_0007_01_000003 (state: COMPLETE, exit status: 1)
16/01/11 15:12:40 INFO yarn.YarnAllocator: Container marked as failed: 
container_1452320323183_0007_01_000003. Exit status: 1. Diagnostics: Exception 
from container-launch.
Container id: container_1452320323183_0007_01_000003
Exit code: 1
Stack trace: ExitCodeException exitCode=1: 
at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
at org.apache.hadoop.util.Shell.run(Shell.java:455)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:715)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:211)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)




Container exited with a non-zero exit code 1

































Reply via email to