[ 
https://issues.apache.org/jira/browse/KYLIN-4522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17119237#comment-17119237
 ] 

Carlos Ignacio Molina López commented on KYLIN-4522:
----------------------------------------------------

Hi again, dear, Xiaoxiang Yu

It seems that the 3.0.2 installation doesn't work as standalone. I was able to 
install and run version 3.0.2 in the same cluster on which I had installed 
previous version 2.5.2, and the error message that happened in version 2.6.6 is 
also happening in version 3.0.2. 

If possible, please share some documentation on how to troubleshoot similar 
errors.

org.apache.kylin.engine.spark.exception.SparkException: OS command error exit 
with return code: 1, error message: 20/05/29 03:01:03 WARN SparkConf: The 
configuration key 'spark.yarn.executor.memoryOverhead' has been deprecated as 
of Spark 2.3 and may be removed in the future. Please use the new key 
'spark.executor.memoryOverhead' instead.
20/05/29 03:01:04 INFO RMProxy: Connecting to ResourceManager at 
ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal/XXX.XXX.XXX.XXX:8032
20/05/29 03:01:05 INFO Client: Requesting a new application from cluster with 4 
NodeManagers
20/05/29 03:01:05 INFO Client: Verifying our application has not requested more 
than the maximum memory capability of the cluster (6144 MB per container)
20/05/29 03:01:05 INFO Client: Will allocate AM container, with 5632 MB memory 
including 512 MB overhead
20/05/29 03:01:05 INFO Client: Setting up container launch context for our AM
20/05/29 03:01:05 INFO Client: Setting up the launch environment for our AM 
container
20/05/29 03:01:05 INFO Client: Preparing resources for our AM container
20/05/29 03:01:06 WARN Client: Neither spark.yarn.jars nor spark.yarn.archive 
is set, falling back to uploading libraries under SPARK_HOME.
20/05/29 03:01:08 INFO Client: Uploading resource 
file:/mnt/tmp/spark-e236db36-0c0d-4d1e-b78a-39dfac3148cf/__spark_libs__1316590902952064148.zip
 -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/__spark_libs__1316590902952064148.zip
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/local/kylin/apache-kylin-3.0.2-bin-hbase1x/lib/kylin-job-3.0.2.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/kylin-job-3.0.2.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-common-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-common-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-server-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-server-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-client-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-client-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-protocol-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-protocol-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-hadoop-compat-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/htrace-core-3.1.0-incubating.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/htrace-core-3.1.0-incubating.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/local/kylin/apache-kylin-3.0.2-bin-hbase1x/tomcat/webapps/kylin/WEB-INF/lib/metrics-core-2.2.0.jar
 -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/metrics-core-2.2.0.jar
20/05/29 03:01:09 WARN Client: Same path resource 
file:///usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar added multiple times to 
distributed cache.
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/usr/lib/hbase/lib/hbase-hadoop2-compat-1.4.7.jar -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hbase-hadoop2-compat-1.4.7.jar
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/etc/spark/conf/hive-site.xml -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/hive-site.xml
20/05/29 03:01:09 INFO Client: Uploading resource 
file:/mnt/tmp/spark-e236db36-0c0d-4d1e-b78a-39dfac3148cf/__spark_conf__645747524997720434.zip
 -> 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590711853580_0024/__spark_conf__.zip
20/05/29 03:01:09 INFO SecurityManager: Changing view acls to: hadoop
20/05/29 03:01:09 INFO SecurityManager: Changing modify acls to: hadoop
20/05/29 03:01:09 INFO SecurityManager: Changing view acls groups to: 
20/05/29 03:01:09 INFO SecurityManager: Changing modify acls groups to: 
20/05/29 03:01:09 INFO SecurityManager: SecurityManager: authentication 
disabled; ui acls disabled; users with view permissions: Set(hadoop); groups 
with view permissions: Set(); users with modify permissions: Set(hadoop); 
groups with modify permissions: Set()
20/05/29 03:01:09 INFO Client: Submitting application 
application_1590711853580_0024 to ResourceManager
20/05/29 03:01:09 INFO YarnClientImpl: Submitted application 
application_1590711853580_0024
20/05/29 03:01:10 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:10 INFO Client: 
client token: N/A
diagnostics: AM container is launched, waiting for AM container to Register 
with RM
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: default
start time: 1590721269736
final status: UNDEFINED
tracking URL: 
http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590711853580_0024/
user: hadoop
20/05/29 03:01:11 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:12 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:13 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:14 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:15 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:15 INFO Client: 
client token: N/A
diagnostics: N/A
ApplicationMaster host: XXX.XXX.XXX.XXX
ApplicationMaster RPC port: 0
queue: default
start time: 1590721269736
final status: UNDEFINED
tracking URL: 
http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590711853580_0024/
user: hadoop
20/05/29 03:01:16 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:17 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:18 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:19 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:20 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:21 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:22 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:23 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:24 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:25 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:26 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:27 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:28 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:29 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:30 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:31 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:32 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:33 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:34 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:35 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:36 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:37 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:38 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:39 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:40 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:41 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:42 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:43 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:44 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:45 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:46 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:47 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:48 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:49 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:50 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:51 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:52 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:53 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:54 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:55 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:56 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:57 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:58 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:01:59 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:01:59 INFO Client: 
client token: N/A
diagnostics: AM container is launched, waiting for AM container to Register 
with RM
ApplicationMaster host: N/A
ApplicationMaster RPC port: -1
queue: default
start time: 1590721269736
final status: UNDEFINED
tracking URL: 
http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590711853580_0024/
user: hadoop
20/05/29 03:02:00 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:02:01 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:02:02 INFO Client: Application report for 
application_1590711853580_0024 (state: ACCEPTED)
20/05/29 03:02:03 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:03 INFO Client: 
client token: N/A
diagnostics: N/A
ApplicationMaster host: XXX.XXX.XXX.XXX
ApplicationMaster RPC port: 0
queue: default
start time: 1590721269736
final status: UNDEFINED
tracking URL: 
http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590711853580_0024/
user: hadoop
20/05/29 03:02:04 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:05 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:06 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:07 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:08 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:09 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:10 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:11 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:12 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:13 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:14 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:15 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:16 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:17 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:18 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:19 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:20 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:21 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:22 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:23 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:24 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:25 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:26 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:27 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:28 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:29 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:30 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:31 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:32 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:33 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:34 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:35 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:36 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:37 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:38 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:39 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:40 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:41 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:42 INFO Client: Application report for 
application_1590711853580_0024 (state: RUNNING)
20/05/29 03:02:43 INFO Client: Application report for 
application_1590711853580_0024 (state: FINISHED)
20/05/29 03:02:43 INFO Client: 
client token: N/A
diagnostics: User class threw exception: java.lang.RuntimeException: error 
execute org.apache.kylin.storage.hbase.steps.SparkCubeHFile. Root cause: Job 
aborted.
at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42)
at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.spark.deploy.yarn.ApplicationMaster$$anon$4.run(ApplicationMaster.scala:721)
Caused by: org.apache.spark.SparkException: Job aborted.
at 
org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:100)
at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1083)
at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1081)
at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1081)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
at 
org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1081)
at 
org.apache.spark.api.java.JavaPairRDD.saveAsNewAPIHadoopDataset(JavaPairRDD.scala:831)
at 
org.apache.kylin.storage.hbase.steps.SparkCubeHFile.execute(SparkCubeHFile.java:238)
at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37)
... 6 more
Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 
Task 1 in stage 1.0 failed 4 times, most recent failure: Lost task 1.3 in stage 
1.0 (TID 15, ip-172-16-3-167.us-west-2.compute.internal, executor 2): 
org.apache.spark.SparkException: Task failed while writing rows
at 
org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:155)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:83)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
at org.apache.spark.scheduler.Task.run(Task.scala:109)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NoClassDefFoundError: Could not initialize class 
org.apache.hadoop.hbase.io.hfile.HFile
at 
org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:880)
at 
org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:805)
at 
org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:739)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.getNewWriter(HFileOutputFormat3.java:224)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:181)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:153)
at 
org.apache.spark.internal.io.HadoopMapReduceWriteConfigUtil.write(SparkHadoopWriter.scala:356)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:130)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:127)
at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1415)
at 
org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:139)
... 8 more

Driver stacktrace:
at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1803)
at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1791)
at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1790)
at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1790)
at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:871)
at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:871)
at scala.Option.foreach(Option.scala:257)
at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:871)
at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2024)
at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1973)
at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1962)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:682)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2034)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2055)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2087)
at 
org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:78)
... 16 more
Caused by: org.apache.spark.SparkException: Task failed while writing rows
at 
org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:155)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:83)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
at org.apache.spark.scheduler.Task.run(Task.scala:109)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NoClassDefFoundError: Could not initialize class 
org.apache.hadoop.hbase.io.hfile.HFile
at 
org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:880)
at 
org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:805)
at 
org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:739)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.getNewWriter(HFileOutputFormat3.java:224)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:181)
at 
org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:153)
at 
org.apache.spark.internal.io.HadoopMapReduceWriteConfigUtil.write(SparkHadoopWriter.scala:356)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:130)
at 
org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:127)
at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1415)
at 
org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:139)
... 8 more

ApplicationMaster host: XXX.XXX.XXX.XXX
ApplicationMaster RPC port: 0
queue: default
start time: 1590721269736
final status: FAILED
tracking URL: 
http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590711853580_0024/
user: hadoop
Exception in thread "main" org.apache.spark.SparkException: Application 
application_1590711853580_0024 finished with failed status
at org.apache.spark.deploy.yarn.Client.run(Client.scala:1165)
at org.apache.spark.deploy.yarn.YarnClusterApplication.start(Client.scala:1520)
at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
20/05/29 03:02:43 INFO ShutdownHookManager: Shutdown hook called
20/05/29 03:02:43 INFO ShutdownHookManager: Deleting directory 
/mnt/tmp/spark-e236db36-0c0d-4d1e-b78a-39dfac3148cf
20/05/29 03:02:43 INFO ShutdownHookManager: Deleting directory 
/mnt/tmp/spark-b44101bb-7107-471d-9bce-920f38590ff8
The command is: 
export HADOOP_CONF_DIR=/etc/hbase/conf && /usr/lib/spark/bin/spark-submit 
--class org.apache.kylin.common.util.SparkEntry --name "Convert Cuboid Data to 
HFile" --conf spark.executor.instances=16 --conf spark.yarn.queue=default 
--conf spark.history.fs.logDirectory=hdfs:///kylin/spark-history --conf 
spark.master=yarn --conf spark.hadoop.yarn.timeline-service.enabled=true --conf 
spark.executor.memory=5G --conf spark.eventLog.enabled=true --conf 
spark.eventLog.dir=hdfs:///kylin/spark-history --conf 
spark.yarn.executor.memoryOverhead=1024 --conf spark.driver.memory=5G --conf 
spark.submit.deployMode=cluster --conf spark.shuffle.service.enabled=true 
--jars 
/usr/lib/hbase/lib/hbase-common-1.4.7.jar,/usr/lib/hbase/lib/hbase-server-1.4.7.jar,/usr/lib/hbase/lib/hbase-client-1.4.7.jar,/usr/lib/hbase/lib/hbase-protocol-1.4.7.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar,/usr/lib/hbase/lib/htrace-core-3.1.0-incubating.jar,/usr/local/kylin/apache-kylin-3.0.2-bin-hbase1x/tomcat/webapps/kylin/WEB-INF/lib/metrics-core-2.2.0.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar,/usr/lib/hbase/lib/hbase-hadoop2-compat-1.4.7.jar,
 /usr/local/kylin/apache-kylin-3.0.2-bin-hbase1x/lib/kylin-job-3.0.2.jar 
-className org.apache.kylin.storage.hbase.steps.SparkCubeHFile -partitions 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/kylin_sales_cube/rowkey_stats/part-r-00000_hfile
 -counterOutput 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/kylin_sales_cube/counter
 -cubename kylin_sales_cube -output 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/kylin_sales_cube/hfile
 -input 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/kylin_sales_cube/cuboid/
 -segmentId c4f3a016-f7bc-7412-563c-e9f043eaee89 -metaUrl 
kylin_metadata@hdfs,path=hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/kylin_sales_cube/metadata
 -hbaseConfPath 
hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-e37b5328-e0b0-37cf-384d-84f1364e9e6c/hbase-conf.xml
at 
org.apache.kylin.engine.spark.SparkExecutable.doWork(SparkExecutable.java:392)
at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:179)
at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)

Thank you and kind regards

Carlos Molina.

> Could not initialize class org.apache.hadoop.hbase.io.hfile.HFile Kylin 2.6.6 
> EMR  5.19
> ---------------------------------------------------------------------------------------
>
>                 Key: KYLIN-4522
>                 URL: https://issues.apache.org/jira/browse/KYLIN-4522
>             Project: Kylin
>          Issue Type: Bug
>          Components: Environment , Job Engine, Others
>    Affects Versions: v2.6.6
>         Environment: Release label: emr-5.19.0
> Hadoop distribution:Amazon 2.8.5
> Applications: Hive 2.3.3, HBase 1.4.7, Spark 2.3.2, Livy 0.5.0, ZooKeeper 
> 3.4.13, Sqoop 1.4.7, Oozie 5.0.0, Pig 0.17.0, HCatalog 2.3.3
>            Reporter: Carlos Ignacio Molina López
>            Priority: Major
>         Attachments: base_2020_05_25_14_29_52.zip
>
>
> Hi,
> I've tried to build the Sample kylin_sales_cube with Spark to run in Amazon 
> EMR Cluster. I saw issue KYLIN-3931 and suggestion is to use the 2.6.6 Engine 
> for Hadoop 3. In EMR Hadoop 3 is only available on EMR 6.0 which is very 
> recent and I had tried to setup versions 2.6.6 and 3.0.2 for Hadoop 3, but in 
> both cases the Kylin Site doesn't show up (Error 404 - Not Found). So I tried 
> to run in EMR 5.19 that has same version of Spark (2.3.2) used in Kylin 2.6.6.
> I am getting "java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.hadoop.hbase.io.hfile.HFile" error message. 
> I had already copied the following jars to Spark Jars folder, as per 
> documentations and what I've read in kylin-issues mailing list archives:
> /usr/lib/hbase/hbase-hadoop-compat-1.4.7.jar
> /usr/lib/hbase/hbase-hadoop2-compat-1.4.7.jar
> /usr/lib/hbase/lib/hbase-common-1.4.7-tests.jar
> /usr/lib/hbase/lib/hbase-common-1.4.7.jar
> /usr/lib/hbase/hbase-client.jar
> /usr/lib/hbase/hbase-client-1.4.7.jar
> /usr/lib/hbase/hbase-server-1.4.7.jar
>  
> This is the output shown on the Step
> {{org.apache.kylin.engine.spark.exception.SparkException: OS command error 
> exit with return code: 1, error message: 20/05/25 14:03:46 WARN SparkConf: 
> The configuration key 'spark.yarn.executor.memoryOverhead' has been 
> deprecated as of Spark 2.3 and may be removed in the future. Please use the 
> new key 'spark.executor.memoryOverhead' 
> instead.org.apache.kylin.engine.spark.exception.SparkException: OS command 
> error exit with return code: 1, error message: 20/05/25 14:03:46 WARN 
> SparkConf: The configuration key 'spark.yarn.executor.memoryOverhead' has 
> been deprecated as of Spark 2.3 and may be removed in the future. Please use 
> the new key 'spark.executor.memoryOverhead' instead.20/05/25 14:03:47 INFO 
> RMProxy: Connecting to ResourceManager at 
> ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal/XXX.XXX.XXX.XXX:803220/05/25 
> 14:03:49 INFO Client: Requesting a new application from cluster with 4 
> NodeManagers20/05/25 14:03:49 INFO Client: Verifying our application has not 
> requested more than the maximum memory capability of the cluster (6144 MB per 
> container)20/05/25 14:03:49 INFO Client: Will allocate AM container, with 
> 5632 MB memory including 512 MB overhead20/05/25 14:03:49 INFO Client: 
> Setting up container launch context for our AM20/05/25 14:03:49 INFO Client: 
> Setting up the launch environment for our AM container20/05/25 14:03:49 INFO 
> Client: Preparing resources for our AM container20/05/25 14:03:51 WARN 
> Client: Neither spark.yarn.jars nor spark.yarn.archive is set, falling back 
> to uploading libraries under SPARK_HOME.20/05/25 14:03:54 INFO Client: 
> Uploading resource 
> file:/mnt/tmp/spark-d26c4f1f-1b8a-4cf8-a05b-842294ce017d/__spark_libs__4034657074333893156.zip
>  -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/__spark_libs__4034657074333893156.zip20/05/25
>  14:03:54 INFO Client: Uploading resource 
> file:/usr/local/kylin/apache-kylin-2.6.6-bin-hbase1x/lib/kylin-job-2.6.6.jar 
> -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/kylin-job-2.6.6.jar20/05/25
>  14:03:55 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-common-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-common-1.4.7.jar20/05/25
>  14:03:55 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-server-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-server-1.4.7.jar20/05/25
>  14:03:55 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-client-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-client-1.4.7.jar20/05/25
>  14:03:55 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-protocol-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-protocol-1.4.7.jar20/05/25
>  14:03:55 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-hadoop-compat-1.4.7.jar20/05/25
>  14:03:56 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/htrace-core-3.1.0-incubating.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/htrace-core-3.1.0-incubating.jar20/05/25
>  14:03:56 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/metrics-core-2.2.0.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/metrics-core-2.2.0.jar20/05/25
>  14:03:56 WARN Client: Same path resource 
> file:///usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar added multiple times 
> to distributed cache.20/05/25 14:03:56 INFO Client: Uploading resource 
> file:/usr/lib/hbase/lib/hbase-hadoop2-compat-1.4.7.jar -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hbase-hadoop2-compat-1.4.7.jar20/05/25
>  14:03:56 INFO Client: Uploading resource file:/etc/spark/conf/hive-site.xml 
> -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/hive-site.xml20/05/25
>  14:03:56 INFO Client: Uploading resource 
> file:/mnt/tmp/spark-d26c4f1f-1b8a-4cf8-a05b-842294ce017d/__spark_conf__1997289269037988671.zip
>  -> 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/user/hadoop/.sparkStaging/application_1590337422418_0043/__spark_conf__.zip20/05/25
>  14:03:56 INFO SecurityManager: Changing view acls to: hadoop20/05/25 
> 14:03:56 INFO SecurityManager: Changing modify acls to: hadoop20/05/25 
> 14:03:56 INFO SecurityManager: Changing view acls groups to: 20/05/25 
> 14:03:56 INFO SecurityManager: Changing modify acls groups to: 20/05/25 
> 14:03:56 INFO SecurityManager: SecurityManager: authentication disabled; ui 
> acls disabled; users  with view permissions: Set(hadoop); groups with view 
> permissions: Set(); users  with modify permissions: Set(hadoop); groups with 
> modify permissions: Set()20/05/25 14:03:56 INFO Client: Submitting 
> application application_1590337422418_0043 to ResourceManager20/05/25 
> 14:03:56 INFO YarnClientImpl: Submitted application 
> application_1590337422418_004320/05/25 14:03:57 INFO Client: Application 
> report for application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:03:57 
> INFO Client:  client token: N/A diagnostics: AM container is launched, 
> waiting for AM container to Register with RM ApplicationMaster host: N/A 
> ApplicationMaster RPC port: -1 queue: default start time: 1590415436952 final 
> status: UNDEFINED tracking URL: 
> http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590337422418_0043/
>  user: hadoop20/05/25 14:03:58 INFO Client: Application report for 
> application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:03:59 INFO 
> Client: Application report for application_1590337422418_0043 (state: 
> ACCEPTED)20/05/25 14:04:00 INFO Client: Application report for 
> application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:04:01 INFO 
> Client: Application report for application_1590337422418_0043 (state: 
> ACCEPTED)20/05/25 14:04:02 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:02 INFO Client: 
>  client token: N/A diagnostics: N/A ApplicationMaster host: XXX.XXX.XXX.XXX 
> ApplicationMaster RPC port: 0 queue: default start time: 1590415436952 final 
> status: UNDEFINED tracking URL: 
> http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590337422418_0043/
>  user: hadoop20/05/25 14:04:03 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:04 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:05 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:06 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:07 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:08 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:09 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:10 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:11 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:12 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:13 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:14 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:15 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:16 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:17 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:18 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:19 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:21 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:22 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:23 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:24 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:25 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:26 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:27 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:28 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:29 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:30 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:31 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:32 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:33 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:34 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:35 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:36 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:37 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:38 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:39 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:40 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:41 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:42 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:43 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:44 INFO Client: Application report for 
> application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:04:44 INFO 
> Client:  client token: N/A diagnostics: AM container is launched, waiting for 
> AM container to Register with RM ApplicationMaster host: N/A 
> ApplicationMaster RPC port: -1 queue: default start time: 1590415436952 final 
> status: UNDEFINED tracking URL: 
> http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590337422418_0043/
>  user: hadoop20/05/25 14:04:45 INFO Client: Application report for 
> application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:04:46 INFO 
> Client: Application report for application_1590337422418_0043 (state: 
> ACCEPTED)20/05/25 14:04:47 INFO Client: Application report for 
> application_1590337422418_0043 (state: ACCEPTED)20/05/25 14:04:48 INFO 
> Client: Application report for application_1590337422418_0043 (state: 
> ACCEPTED)20/05/25 14:04:49 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:49 INFO Client: 
>  client token: N/A diagnostics: N/A ApplicationMaster host: XXX.XXX.XXX.XXX 
> ApplicationMaster RPC port: 0 queue: default start time: 1590415436952 final 
> status: UNDEFINED tracking URL: 
> http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590337422418_0043/
>  user: hadoop20/05/25 14:04:50 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:51 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:52 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:53 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:54 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:55 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:56 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:57 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:04:58 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:04:59 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:00 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:01 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:02 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:03 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:04 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:05 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:06 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:07 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:08 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:09 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:10 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:11 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:12 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:13 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:14 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:15 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:16 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:17 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:18 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:19 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:20 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:21 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:22 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:23 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:24 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:25 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> RUNNING)20/05/25 14:05:26 INFO Client: Application report for 
> application_1590337422418_0043 (state: RUNNING)20/05/25 14:05:27 INFO Client: 
> Application report for application_1590337422418_0043 (state: 
> FINISHED)20/05/25 14:05:27 INFO Client:  client token: N/A diagnostics: User 
> class threw exception: java.lang.RuntimeException: error execute 
> org.apache.kylin.storage.hbase.steps.SparkCubeHFile. Root cause: Job aborted. 
> at 
> org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42)
>  at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44) at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$4.run(ApplicationMaster.scala:721)Caused
>  by: org.apache.spark.SparkException: Job aborted. at 
> org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:100)
>  at 
> org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1083)
>  at 
> org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1081)
>  at 
> org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:1081)
>  at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>  at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>  at org.apache.spark.rdd.RDD.withScope(RDD.scala:363) at 
> org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:1081)
>  at 
> org.apache.spark.api.java.JavaPairRDD.saveAsNewAPIHadoopDataset(JavaPairRDD.scala:831)
>  at 
> org.apache.kylin.storage.hbase.steps.SparkCubeHFile.execute(SparkCubeHFile.java:238)
>  at 
> org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37)
>  ... 6 moreCaused by: org.apache.spark.SparkException: Job aborted due to 
> stage failure: Task 1 in stage 1.0 failed 4 times, most recent failure: Lost 
> task 1.3 in stage 1.0 (TID 15, ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal, 
> executor 3): org.apache.spark.SparkException: Task failed while writing rows 
> at 
> org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:155)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:83)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) at 
> org.apache.spark.scheduler.Task.run(Task.scala:109) at 
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)Caused by: 
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.hadoop.hbase.io.hfile.HFile at 
> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:880)
>  at 
> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:805)
>  at 
> org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:739)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.getNewWriter(HFileOutputFormat3.java:224)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:181)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:153)
>  at 
> org.apache.spark.internal.io.HadoopMapReduceWriteConfigUtil.write(SparkHadoopWriter.scala:356)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:130)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:127)
>  at 
> org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1415)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:139)
>  ... 8 more}}
> {{Driver stacktrace: at 
> org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1803)
>  at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1791)
>  at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1790)
>  at 
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
>  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at 
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1790) 
> at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:871)
>  at 
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:871)
>  at scala.Option.foreach(Option.scala:257) at 
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:871)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2024)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1973)
>  at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1962)
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) at 
> org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:682) at 
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2034) at 
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2055) at 
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2087) at 
> org.apache.spark.internal.io.SparkHadoopWriter$.write(SparkHadoopWriter.scala:78)
>  ... 16 moreCaused by: org.apache.spark.SparkException: Task failed while 
> writing rows at 
> org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:155)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:83)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$3.apply(SparkHadoopWriter.scala:78)
>  at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87) at 
> org.apache.spark.scheduler.Task.run(Task.scala:109) at 
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)Caused by: 
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.hadoop.hbase.io.hfile.HFile at 
> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:880)
>  at 
> org.apache.hadoop.hbase.regionserver.StoreFile$Writer.<init>(StoreFile.java:805)
>  at 
> org.apache.hadoop.hbase.regionserver.StoreFile$WriterBuilder.build(StoreFile.java:739)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.getNewWriter(HFileOutputFormat3.java:224)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:181)
>  at 
> org.apache.kylin.storage.hbase.steps.HFileOutputFormat3$1.write(HFileOutputFormat3.java:153)
>  at 
> org.apache.spark.internal.io.HadoopMapReduceWriteConfigUtil.write(SparkHadoopWriter.scala:356)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:130)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:127)
>  at 
> org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1415)
>  at 
> org.apache.spark.internal.io.SparkHadoopWriter$.org$apache$spark$internal$io$SparkHadoopWriter$$executeTask(SparkHadoopWriter.scala:139)
>  ... 8 more}}
> {{ ApplicationMaster host: XXX.XXX.XXX.XXX ApplicationMaster RPC port: 0 
> queue: default start time: 1590415436952 final status: FAILED tracking URL: 
> http://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:20888/proxy/application_1590337422418_0043/
>  user: hadoopException in thread "main" org.apache.spark.SparkException: 
> Application application_1590337422418_0043 finished with failed status at 
> org.apache.spark.deploy.yarn.Client.run(Client.scala:1165) at 
> org.apache.spark.deploy.yarn.YarnClusterApplication.start(Client.scala:1520) 
> at 
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:894)
>  at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198) 
> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228) at 
> org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at 
> org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)20/05/25 14:05:27 
> INFO ShutdownHookManager: Shutdown hook called20/05/25 14:05:27 INFO 
> ShutdownHookManager: Deleting directory 
> /mnt/tmp/spark-04e9eed4-d16e-406c-9fb0-972cf355db0920/05/25 14:05:27 INFO 
> ShutdownHookManager: Deleting directory 
> /mnt/tmp/spark-d26c4f1f-1b8a-4cf8-a05b-842294ce017dThe command is: export 
> HADOOP_CONF_DIR=/etc/hadoop/conf && /usr/lib/spark/bin/spark-submit --class 
> org.apache.kylin.common.util.SparkEntry  --conf spark.executor.instances=40  
> --conf spark.yarn.queue=default  --conf 
> spark.history.fs.logDirectory=hdfs:///kylin/spark-history  --conf 
> spark.master=yarn  --conf spark.hadoop.yarn.timeline-service.enabled=false  
> --conf spark.executor.memory=5G  --conf spark.eventLog.enabled=true  --conf 
> spark.eventLog.dir=hdfs:///kylin/spark-history  --conf 
> spark.yarn.executor.memoryOverhead=1024  --conf spark.driver.memory=5G  
> --conf spark.submit.deployMode=cluster  --conf 
> spark.shuffle.service.enabled=true --jars 
> /usr/lib/hbase/lib/hbase-common-1.4.7.jar,/usr/lib/hbase/lib/hbase-server-1.4.7.jar,/usr/lib/hbase/lib/hbase-client-1.4.7.jar,/usr/lib/hbase/lib/hbase-protocol-1.4.7.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar,/usr/lib/hbase/lib/htrace-core-3.1.0-incubating.jar,/usr/lib/hbase/lib/metrics-core-2.2.0.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.7.jar,/usr/lib/hbase/lib/hbase-hadoop2-compat-1.4.7.jar,
>  /usr/local/kylin/apache-kylin-2.6.6-bin-hbase1x/lib/kylin-job-2.6.6.jar 
> -className org.apache.kylin.storage.hbase.steps.SparkCubeHFile -partitions 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/kylin_sales_cube/rowkey_stats/part-r-00000_hfile
>  -counterOutput 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/kylin_sales_cube/counter
>  -cubename kylin_sales_cube -output 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/kylin_sales_cube/hfile
>  -input 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/kylin_sales_cube/cuboid/
>  -segmentId 0d22a9ac-5256-02cd-a5b9-44de5247871f -metaUrl 
> kylin_metadata@hdfs,path=hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/kylin_sales_cube/metadata
>  -hbaseConfPath 
> hdfs://ip-XXX-XXX-XXX-XXX.us-west-2.compute.internal:8020/kylin/kylin_metadata/kylin-b75c7f69-2ebf-c5c3-4a6e-b01f177d911f/hbase-conf.xml
>  at 
> org.apache.kylin.engine.spark.SparkExecutable.doWork(SparkExecutable.java:347)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:167)
>  at 
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:71)
>  at 
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:167)
>  at 
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:114)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)}}
>  
> {{Please suggest how this issue can be troubleshooted.}}
> Thank you and kind regards
> {{Carlos Molina.}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to