Hi,
here is my configuration in kylin.properties
 
kylin.env.hadoop-conf-dir=/usr/local/kylin/hadoop-conf

[kylin@gateway conf]$ ll /usr/local/kylin/hadoop-conf
总用量 20
lrwxrwxrwx 1 kylin kylin    42 6月  28 13:31 core-site.xml -> 
/usr/local/hadoop/etc/hadoop/core-site.xml
lrwxrwxrwx 1 kylin kylin    36 6月  28 13:32 hbase-site.xml -> 
/usr/local/hbase/conf/hbase-site.xml
lrwxrwxrwx 1 kylin kylin    42 6月  28 13:31 hdfs-site.xml -> 
/usr/local/hadoop/etc/hadoop/hdfs-site.xml
-rw-r--r-- 1 kylin kylin 17924 6月  28 13:33 hive-site.xml
lrwxrwxrwx 1 kylin kylin    42 6月  28 13:31 yarn-site.xml -> 
/usr/local/hadoop/etc/hadoop/yarn-site.xml

and all log is below 

OS command error exit with 1 -- export 
HADOOP_CONF_DIR=/usr/local/kylin/hadoop-conf && 
/usr/local/spark/bin/spark-submit --class 
org.apache.kylin.common.util.SparkEntry  --conf spark.executor.instances=1  
--conf 
spark.yarn.jar=hdfs://ns1/kylin/spark/spark-assembly-1.6.0-cdh5.9.0-hadoop2.6.0-cdh5.9.0.jar
  --conf spark.yarn.queue=default  --conf 
spark.history.fs.logDirectory=hdfs:///kylin/spark-history  --conf 
spark.master=yarn  --conf spark.executor.memory=4G  --conf 
spark.eventLog.enabled=true  --conf 
spark.eventLog.dir=hdfs:///kylin/spark-history  --conf spark.executor.cores=2  
--conf spark.submit.deployMode=client --files 
/usr/local/hbase/conf/hbase-site.xml --jars 
/usr/local/hbase/lib/htrace-core-3.2.0-incubating.jar,/usr/local/hbase/lib/hbase-client-1.2.0-cdh5.9.0.jar,/usr/local/hbase/lib/hbase-common-1.2.0-cdh5.9.0.jar,/usr/local/hbase/lib/hbase-protocol-1.2.0-cdh5.9.0.jar,/usr/local/hbase/lib/metrics-core-2.2.0.jar,/usr/local/hbase/lib/guava-12.0.1.jar,
 /usr/local/kylin/lib/kylin-job-2.0.0.jar -className 
org.apache.kylin.engine.spark.SparkCubingByLayer -hiveTable 
default.kylin_intermediate_kylin_sales_cube_d7955f9a_d290_4479_866c_5745f7880b81
 -output 
hdfs:///kylin/kylin_metadata/kylin-09df3734-6dcf-41b2-99a1-02e2b4e16f9f/kylin_sales_cube/cuboid/
 -segmentId d7955f9a-d290-4479-866c-5745f7880b81 -confPath 
/usr/local/kylin/conf -cubename kylin_sales_cube
SparkEntry args:-className org.apache.kylin.engine.spark.SparkCubingByLayer 
-hiveTable 
default.kylin_intermediate_kylin_sales_cube_d7955f9a_d290_4479_866c_5745f7880b81
 -output 
hdfs:///kylin/kylin_metadata/kylin-09df3734-6dcf-41b2-99a1-02e2b4e16f9f/kylin_sales_cube/cuboid/
 -segmentId d7955f9a-d290-4479-866c-5745f7880b81 -confPath 
/usr/local/kylin/conf -cubename kylin_sales_cube
Abstract Application args:-hiveTable 
default.kylin_intermediate_kylin_sales_cube_d7955f9a_d290_4479_866c_5745f7880b81
 -output 
hdfs:///kylin/kylin_metadata/kylin-09df3734-6dcf-41b2-99a1-02e2b4e16f9f/kylin_sales_cube/cuboid/
 -segmentId d7955f9a-d290-4479-866c-5745f7880b81 -confPath 
/usr/local/kylin/conf -cubename kylin_sales_cube
spark.yarn.driver.memoryOverhead is set but does not apply in client mode.
spark.driver.cores is set but does not apply in client mode.
Hive history 
file=/usr/local/hive2/logs/kylin/hive_job_log_fe557857-7a8e-407f-aeab-94044b59931d_2113658910.txt
Hive history 
file=/usr/local/hive2/logs/kylin/hive_job_log_df0eaa5c-b047-43a3-bd07-0cc8c0ec148c_2093334660.txt
Exception in thread "main" java.lang.RuntimeException: error execute 
org.apache.kylin.engine.spark.SparkCubingByLayer
        at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42)
        at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
        at 
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.IllegalArgumentException: Failed to find metadata store by 
url: kylin_metadata@hbase
        at 
org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99)
        at 
org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110)
        at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:812)
        at 
org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:732)
        at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:143)
        at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:107)
        at 
org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160)
        at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37)
        ... 10 more

Thank you very much!



[email protected]
 
From: ShaoFeng Shi
Date: 2017-06-28 19:49
To: user
Subject: Re: Failed to find metadata store by url: kylin_metadata@hbase
Hi, 

Could you pls provide the full log? I'm afraid the root cause wasn't included 
here. There should be other error before or after this. Please also check 
whether the hbase-site.xml is in the HADOOP_CONF folder and Spark executors can 
connect with HBase cluster.

Here is a similar question before: 
http://apache-kylin.74782.x6.nabble.com/A-problem-in-cube-building-with-SPARK-td8009.html

2017-06-28 16:58 GMT+08:00 [email protected] <[email protected]>:
Hi,
    I upgrade Kylin to 2.0.
   I can successs run at Mapreduce engine now ,but when i build a cube using 
spark cube, it shows errors below.
How could I solve it. Thanks 


Exception in thread "main" java.lang.RuntimeException: error execute 
org.apache.kylin.engine.spark.SparkCubingByLayer
at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:42)
at org.apache.kylin.common.util.SparkEntry.main(SparkEntry.java:44)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.IllegalArgumentException: Failed to find metadata store by 
url: kylin_metadata@hbase
at 
org.apache.kylin.common.persistence.ResourceStore.createResourceStore(ResourceStore.java:99)
at 
org.apache.kylin.common.persistence.ResourceStore.getStore(ResourceStore.java:110)
at org.apache.kylin.cube.CubeManager.getStore(CubeManager.java:812)
at org.apache.kylin.cube.CubeManager.loadAllCubeInstance(CubeManager.java:732)
at org.apache.kylin.cube.CubeManager.<init>(CubeManager.java:143)
at org.apache.kylin.cube.CubeManager.getInstance(CubeManager.java:107)
at 
org.apache.kylin.engine.spark.SparkCubingByLayer.execute(SparkCubingByLayer.java:160)
at 
org.apache.kylin.common.util.AbstractApplication.execute(AbstractApplication.java:37)
... 10 more



[email protected]



-- 
Best regards,

Shaofeng Shi 史少锋

Reply via email to