There are two cubes using the same dimension table, and when both cubes are 
built at the same time, both will do delete and create operations, causing one 
of them to fail, Please see the log
pool-9-thread-10 and pool-9-thread-9 are two cube builds that are executed 
separately, pool-9-thread-10 failed


2019-07-09 15:49:03,870 INFO [<b>pool-9-thread-10</b>] 
common.ShellExecutable:52 : executing:hive -e "USE kylin_flat_db;
SET dfs.replication=2;
SET hive.exec.compress.output=true;
SET hive.auto.convert.join.noconditionaltask=true;
SET hive.auto.convert.join.noconditionaltask.size=100000000;
SET mapreduce.output.fileoutputformat.compress.type=BLOCK;
SET mapreduce.job.split.metainfo.maxsize=-1;


DROP TABLE IF EXISTS 
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
CREATE TABLE IF NOT EXISTS 
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
LOCATION 
'/kylin/kylin_metadata/kylin-88e0f1f2-c55e-44a2-88a9-37be8e07f445/kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW'
AS SELECT * FROM EDW.D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;


"






2019-07-09 15:49:53,997 INFO [<b>pool-9-thread-9</b>] common.ShellExecutable:52 
: executing:hive -e "USE kylin_flat_db;
SET dfs.replication=2;
SET hive.exec.compress.output=true;
SET hive.auto.convert.join.noconditionaltask=true;
SET hive.auto.convert.join.noconditionaltask.size=100000000;
SET mapreduce.output.fileoutputformat.compress.type=BLOCK;
SET mapreduce.job.split.metainfo.maxsize=-1;


DROP TABLE IF EXISTS 
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
CREATE TABLE IF NOT EXISTS 
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
LOCATION 
'/kylin/kylin_metadata/kylin-feb1f455-3fa1-4d14-ba7c-b5d5fb9969b4/kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW'
AS SELECT * FROM EDW.D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;


"


2019-07-09 15:50:07,991 ERROR [pool-9-thread-10] 
common.HadoopShellExecutable:65 : error execute 
HadoopShellExecutable{id=88e0f1f2-c55e-44a2-88a9-37be8e07f445-04, name=Build 
Dimension Dicti
onary, state=RUNNING}
java.lang.RuntimeException: cannot get HiveTableMeta
        at org.apache.kylin.source.hive.HiveTable.<init>(HiveTable.java:49)
        at 
org.apache.kylin.source.hive.HiveSource.createReadableTable(HiveSource.java:49)
        at 
org.apache.kylin.source.SourceFactory.createReadableTable(SourceFactory.java:42)
        at 
org.apache.kylin.cube.CubeManager.buildSnapshotTable(CubeManager.java:269)
        at 
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:58)
        at 
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:41)
        at 
org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:54)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
        at 
org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
        at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113)
        at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:57)
        at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113)
        at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:136)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
Caused by: 
NoSuchObjectException(message:KYLIN_FLAT_DB.kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
 table not found)
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result$get_table_resultStandardScheme.read(ThriftHiveMetastore.java:35066)
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result$get_table_resultStandardScheme.read(ThriftHiveMetastore.java:35034)
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result.read(ThriftHiveMetastore.java:34965)
        at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_table(ThriftHiveMetastore.java:1218)
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table(ThriftHiveMetastore.java:1204)
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1209)
        at 
org.apache.kylin.source.hive.CLIHiveClient.getHiveTableMeta(CLIHiveClient.java:79)
        at org.apache.kylin.source.hive.HiveTable.<init>(HiveTable.java:47)
        ... 16 more

Reply via email to