Hi,
This looks caused by class load conflict with jar provided by env. Here is
a list of Hadoop Platform we verified and test :
https://cwiki.apache.org/confluence/display/KYLIN/Support+Hadoop+Version+Matrix+of+Kylin+4.0.0
.
Please let us know your env information(version), or you can choose to use our
recommended hadoop platform.
--
Best wishes to you !
From :Xiaoxiang Yu
在 2021-08-25 15:38:25,"washyou112" <[email protected]> 写道:
Kylin 4.0 error in the first step of building cube
kylin.log:
2021-08-24 10:45:00,232 ERROR [Thread-1] application.JobMonitor : Job failed
the 1 times.
java.lang.NoSuchMethodError:
com.fasterxml.jackson.databind.JsonMappingException.<init>(Ljava/io/Closeable;Ljava/lang/String;)V
at
com.fasterxml.jackson.module.scala.JacksonModule$class.setupModule(JacksonModule.scala:61)
at
com.fasterxml.jackson.module.scala.DefaultScalaModule.setupModule(DefaultScalaModule.scala:17)
at
com.fasterxml.jackson.databind.ObjectMapper.registerModule(ObjectMapper.java:718)
at org.apache.spark.util.JsonProtocol$.<init>(JsonProtocol.scala:60)
at org.apache.spark.util.JsonProtocol$.<clinit>(JsonProtocol.scala)
at
org.apache.spark.scheduler.EventLoggingListener$.initEventLog(EventLoggingListener.scala:353)
at
org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:135)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:533)
at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2549)
at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:944)
at
org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:935)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:935)
at
org.apache.kylin.engine.spark.application.SparkApplication.execute(SparkApplication.java:283)
at
org.apache.kylin.engine.spark.application.SparkApplication.execute(SparkApplication.java:89)
at org.apache.spark.application.JobWorker$$anon$2.run(JobWorker.scala:55)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2021-08-24 10:45:00,246 ERROR [Thread-1] application.JobWorkSpace : Job failed
eventually. Reason: Error occurred when generate retry configuration.
java.util.NoSuchElementException: spark.executor.memory
at org.apache.spark.SparkConf$$anonfun$get$1.apply(SparkConf.scala:245)
at org.apache.spark.SparkConf$$anonfun$get$1.apply(SparkConf.scala:245)
at scala.Option.getOrElse(Option.scala:121)
at org.apache.spark.SparkConf.get(SparkConf.scala:245)
at
org.apache.spark.autoheal.ExceptionTerminator$.incMemory(ExceptionTerminator.scala:70)
at
org.apache.spark.autoheal.ExceptionTerminator$.resolveException(ExceptionTerminator.scala:45)
at
org.apache.spark.application.JobMonitor.handleResourceLack(JobMonitor.scala:53)
at
org.apache.spark.application.JobMonitor$$anon$1.onReceive(JobMonitor.scala:33)
at
org.apache.spark.scheduler.KylinJobEventLoop$$anonfun$onReceive$1.apply(KylinJobEventLoop.scala:42)
at
org.apache.spark.scheduler.KylinJobEventLoop$$anonfun$onReceive$1.apply(KylinJobEventLoop.scala:42)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
at
org.apache.spark.scheduler.KylinJobEventLoop.onReceive(KylinJobEventLoop.scala:42)
at
org.apache.spark.scheduler.KylinJobEventLoop.onReceive(KylinJobEventLoop.scala:29)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
Is this an jar package conflict problem? How to solve it?
| |
washyou112
|
|
[email protected]
|
签名由网易邮箱大师定制