Hi, 
    It seems the problem is following 
    "60505 [dispatcher-event-loop-6] ERROR  
org.apache.spark.scheduler.cluster.YarnScheduler  - Lost executor 1 on 
*********: Container marked as failed:"
It usually comes out with not enough memory for your yarn so that yarn 
container is closed because of lack of memory , you can go to yarn resource 
manager web page to see more details with yarn log. 
        If it's the memory issue, you can try to allocate more memory for spark 
yarn executor by change the following configuration item in 
"$KYLIN_HOME/conf/kylin.properties"
    kylin.engine.spark-conf.spark.yarn.executor.memoryOverhead=384
   

-------------------
Best wishes,
Rupeng Wang
 

在 2019/8/29 14:57,“Gourav Gupta”<[email protected]> 写入:

    Hi Sir,
    
    I have installed and configured Apache Kylin 2.4 on Cloudera Platform for
    creating the Cube.
    
    I have been able to create a cube in MapReduce mode but getting the
    below-mentioned caveat while executes on spark mode. I have had followed
    all the steps and tried many remedies for debugging the problem.
    
    
    
    Please let me know how to resolve this bug. Thanks in Advance.
    
    
    
    
    
    1091 [main] ERROR org.apache.spark.SparkContext  - Error adding jar
    (java.lang.IllegalArgumentException: requirement failed: JAR
    kylin-job-2.4.0.jar already registered.), was the --addJars option used?
    
    [Stage 0:>                                                          (0 + 0)
    / 2]
    [Stage 0:>                                                          (0 + 2)
    / 2]
    
    
    60505 [dispatcher-event-loop-6] ERROR
    org.apache.spark.scheduler.cluster.YarnScheduler  - Lost executor 1 on **
    *******: Container marked as failed:
    container_e62_1566915974858_6628_01_000003 on host: *******. Exit status:
    50. Diagnostics: Exception from container-launch.
    Container id: container_e62_1566915974858_6628_01_000003
    Exit code: 50
    Stack trace: ExitCodeException exitCode=50:
    at org.apache.hadoop.util.Shell.runCommand(Shell.java:604)
    at org.apache.hadoop.util.Shell.run(Shell.java:507)
    at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:789)
    at
    
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:213)
    at
    
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:302)
    at
    
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:82)
    at java.util.concurrent.FutureTask.run(FutureTask.java:262)
    at
    
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
    at
    
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
    at java.lang.Thread.run(Thread.java:745)
    
    
    Container exited with a non-zero exit code 50
    
    82664 [dispatcher-event-loop-5] ERROR
    org.apache.spark.scheduler.cluster.YarnScheduler
     - Lost executor 2 on *******: Container marked as failed:
    container_e62_1566915974858_6628_01_000004 on host: *******. Exit status:
    50. Diagnostics: Exception from container-launch.
    Container id: container_e62_1566915974858_6628_01_000004
    Exit code: 50
    Stack trace: ExitCodeException exitCode=50:
    at org.apache.hadoop.util.Shell.runCommand(Shell.java:604)
    at org.apache.hadoop.util.Shell.run(Shell.java:507)
    at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:789)
    at org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.
    launchContainer(DefaultContainerExecutor.java:213)
    at org.apache.hadoop.yarn.server.nodemanager.containermanager.
    launcher.ContainerLaunch.call(ContainerLaunch.java:302)
    at org.apache.hadoop.yarn.server.nodemanager.containermanager.
    launcher.ContainerLaunch.call(ContainerLaunch.java:82)
    at java.util.concurrent.FutureTask.run(FutureTask.java:262)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(
    ThreadPoolExecutor.java:1145)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(
    ThreadPoolExecutor.java:615)
    at java.lang.Thread.run(Thread.java:745)
    
    
    Container exited with a non-zero exit code 50
    
    
    The command is:
    export HADOOP_CONF_DIR=/etc/hadoop/conf && /usr/lib/spark/bin/spark-submit
    --class org.apache.kylin.common.util.SparkEntry  --conf
    spark.executor.instances=1  --conf spark.yarn.archive=hdfs://
    namenode:8020/kylin/spark/spark-libs.jar  --conf spark.yarn.queue=default
     --conf spark.yarn.am.extraJavaOptions=-Dhdp.version=current  --conf
    spark.history.fs.logDirectory=hdfs:///kylin/spark-history  --conf
    spark.driver.extraJavaOptions=-Dhdp.version=current  --conf
    spark.io.compression.codec=org.apache.spark.io.SnappyCompressionCodec
     --conf spark.master=yarn  --conf
    spark.executor.extraJavaOptions=-Dhdp.version=current
     --conf spark.hadoop.yarn.timeline-service.enabled=false  --conf
    spark.executor.memory=4G  --conf spark.eventLog.enabled=true  --conf
    spark.eventLog.dir=hdfs:///kylin/spark-history  --conf
    spark.executor.cores=2  --conf spark.submit.deployMode=cluster --jars
    /opt/apache-kylin-2.4.0-bin-cdh57/lib/kylin-job-2.4.0.jar
    /opt/apache-kylin-2.4.0-bin-cdh57/lib/kylin-job-2.4.0.jar -className
    org.apache.kylin.engine.spark.SparkCubingByLayer -hiveTable
    
default.kylin_intermediate_kylin_sales_cube_c1526d16_9719_4dec_be41_346f43654e02
    -input hdfs://nameservice1/kylin/kylin_metadata/kylin-2159d40b-
    f14e-4500-af95-1fbfd5a4073f/kylin_intermediate_kylin_
    sales_cube_c1526d16_9719_4dec_be41_346f43654e02 -segmentId
    c1526d16-9719-4dec-be41-346f43654e02 -metaUrl kylin_metadata@hdfs,path=hdfs:
    //nameservice1/kylin/kylin_metadata/kylin-2159d40b-f14e-
    4500-af95-1fbfd5a4073f/kylin_sales_cube/metadata -output
    hdfs://nameservice1/kylin/kylin_metadata/kylin-2159d40b-
    f14e-4500-af95-1fbfd5a4073f/kylin_sales_cube/cuboid/ -cubename
    kylin_sales_cube
    

Reply via email to