There are two cubes using the same dimension table, and when both cubes are
built at the same time, both will do delete and create operations, causing one
of them to fail, Please see the log
pool-9-thread-10 and pool-9-thread-9 are two cube builds that are executed
separately, pool-9-thread-10 failed
2019-07-09 15:49:03,870 INFO [<b>pool-9-thread-10</b>]
common.ShellExecutable:52 : executing:hive -e "USE kylin_flat_db;
SET dfs.replication=2;
SET hive.exec.compress.output=true;
SET hive.auto.convert.join.noconditionaltask=true;
SET hive.auto.convert.join.noconditionaltask.size=100000000;
SET mapreduce.output.fileoutputformat.compress.type=BLOCK;
SET mapreduce.job.split.metainfo.maxsize=-1;
DROP TABLE IF EXISTS
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
CREATE TABLE IF NOT EXISTS
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
LOCATION
'/kylin/kylin_metadata/kylin-88e0f1f2-c55e-44a2-88a9-37be8e07f445/kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW'
AS SELECT * FROM EDW.D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
"
2019-07-09 15:49:53,997 INFO [<b>pool-9-thread-9</b>] common.ShellExecutable:52
: executing:hive -e "USE kylin_flat_db;
SET dfs.replication=2;
SET hive.exec.compress.output=true;
SET hive.auto.convert.join.noconditionaltask=true;
SET hive.auto.convert.join.noconditionaltask.size=100000000;
SET mapreduce.output.fileoutputformat.compress.type=BLOCK;
SET mapreduce.job.split.metainfo.maxsize=-1;
DROP TABLE IF EXISTS
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
CREATE TABLE IF NOT EXISTS
kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
LOCATION
'/kylin/kylin_metadata/kylin-feb1f455-3fa1-4d14-ba7c-b5d5fb9969b4/kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW'
AS SELECT * FROM EDW.D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW;
"
2019-07-09 15:50:07,991 ERROR [pool-9-thread-10]
common.HadoopShellExecutable:65 : error execute
HadoopShellExecutable{id=88e0f1f2-c55e-44a2-88a9-37be8e07f445-04, name=Build
Dimension Dicti
onary, state=RUNNING}
java.lang.RuntimeException: cannot get HiveTableMeta
at org.apache.kylin.source.hive.HiveTable.<init>(HiveTable.java:49)
at
org.apache.kylin.source.hive.HiveSource.createReadableTable(HiveSource.java:49)
at
org.apache.kylin.source.SourceFactory.createReadableTable(SourceFactory.java:42)
at
org.apache.kylin.cube.CubeManager.buildSnapshotTable(CubeManager.java:269)
at
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:58)
at
org.apache.kylin.cube.cli.DictionaryGeneratorCLI.processSegment(DictionaryGeneratorCLI.java:41)
at
org.apache.kylin.engine.mr.steps.CreateDictionaryJob.run(CreateDictionaryJob.java:54)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at
org.apache.kylin.engine.mr.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:63)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113)
at
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:57)
at
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:113)
at
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:136)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by:
NoSuchObjectException(message:KYLIN_FLAT_DB.kylin_intermediate_EDW_D_EDW_HOTEL_PROFILE_INFO_FIN_ADD_INDEX_VIEW
table not found)
at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result$get_table_resultStandardScheme.read(ThriftHiveMetastore.java:35066)
at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result$get_table_resultStandardScheme.read(ThriftHiveMetastore.java:35034)
at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_table_result.read(ThriftHiveMetastore.java:34965)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:78)
at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_table(ThriftHiveMetastore.java:1218)
at
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table(ThriftHiveMetastore.java:1204)
at
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1209)
at
org.apache.kylin.source.hive.CLIHiveClient.getHiveTableMeta(CLIHiveClient.java:79)
at org.apache.kylin.source.hive.HiveTable.<init>(HiveTable.java:47)
... 16 more