What's your Kylin version? One possible case is, hive-site.xml is missing from Spark's context, then it couldn't find the intermediate table. Please make sure hive-site.xml is appeared in the HADOOP_CONF_DIR, or you can create a symbolic to if in KYLIN_HOME/spark/conf folder.
2018-02-06 6:00 GMT+08:00 Ge Silas <[email protected]>: > Hi, > > Did the previous building steps successfully create the Hive intermediate > tables in Hive? > > Thanks, > Silas > > On 30 Jan 2018, at 3:09 AM, macwork <[email protected]> wrote: > > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in [jar:file:/home/hadoop/tmp/nm- > local-dir/filecache/13/spark-libs.jar/slf4j-log4j12-1.7.16. > jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/home/hadoop/ > software/hadoop-2.7.3/share/hadoop/common/lib/slf4j- > log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an > explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > 18/01/30 19:05:28 INFO util.SignalUtils: Registered signal handler for TERM > 18/01/30 19:05:28 INFO util.SignalUtils: Registered signal handler for HUP > 18/01/30 19:05:28 INFO util.SignalUtils: Registered signal handler for INT > 18/01/30 19:05:29 INFO yarn.ApplicationMaster: Preparing Local resources > 18/01/30 19:05:30 INFO yarn.ApplicationMaster: ApplicationAttemptId: > appattempt_1516010594436_0185_000002 > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing view acls to: hadoop > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing modify acls to: > hadoop > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing view acls groups to: > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing modify acls groups > to: > 18/01/30 19:05:30 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(hadoop); groups with view permissions: Set(); users with modify > permissions: Set(hadoop); groups with modify permissions: Set() > 18/01/30 19:05:30 INFO yarn.ApplicationMaster: Starting the user > application in a separate Thread > 18/01/30 19:05:30 INFO yarn.ApplicationMaster: Waiting for spark context > initialization... > 18/01/30 19:05:30 INFO spark.SparkContext: Running Spark version 2.1.2 > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing view acls to: hadoop > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing modify acls to: > hadoop > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing view acls groups to: > 18/01/30 19:05:30 INFO spark.SecurityManager: Changing modify acls groups > to: > 18/01/30 19:05:30 INFO spark.SecurityManager: SecurityManager: > authentication disabled; ui acls disabled; users with view permissions: > Set(hadoop); groups with view permissions: Set(); users with modify > permissions: Set(hadoop); groups with modify permissions: Set() > 18/01/30 19:05:31 INFO util.Utils: Successfully started service > 'sparkDriver' on port 42339. > 18/01/30 19:05:31 INFO spark.SparkEnv: Registering MapOutputTracker > 18/01/30 19:05:31 INFO spark.SparkEnv: Registering BlockManagerMaster > 18/01/30 19:05:31 INFO storage.BlockManagerMasterEndpoint: Using > org.apache.spark.storage.DefaultTopologyMapper for getting topology > information > 18/01/30 19:05:31 INFO storage.BlockManagerMasterEndpoint: > BlockManagerMasterEndpoint up > 18/01/30 19:05:31 INFO storage.DiskBlockManager: Created local directory > at /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/ > application_1516010594436_0185/blockmgr-e9044b90-ff87- > 4ad3-9ae7-cf85a131ba4e > 18/01/30 19:05:31 INFO memory.MemoryStore: MemoryStore started with > capacity 366.3 MB > 18/01/30 19:05:31 INFO spark.SparkEnv: Registering OutputCommitCoordinator > 18/01/30 19:05:31 INFO util.log: Logging initialized @3601ms > 18/01/30 19:05:31 INFO ui.JettyUtils: Adding filter: > org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter > 18/01/30 19:05:31 INFO server.Server: jetty-9.2.z-SNAPSHOT > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@1727271{/jobs,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@5a8c27d3{/jobs/json,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@201b73c{/jobs/job,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6dc9732e{/jobs/job/json,null, > AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@7f3a751c{/stages,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@44336797{/stages/json,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@4b84807a{/stages/stage,null, > AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6d72da27{/stages/stage/json, > null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@41ccc000{/stages/pool,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@1e1c0871{/stages/pool/json, > null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@71f4dc98{/storage,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@a2c9363{/storage/json,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@2d479ee6{/storage/rdd,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@53e654bf{/storage/rdd/json, > null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@4215247{/environment,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6af87e2b{/environment/json, > null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@7bd97d3c{/executors,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@1afb9d54{/executors/json,null, > AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@14c64726{/executors/ > threadDump,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@3de67ab3{/executors/threadDump/json,null, > AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@d66c0ae{/static,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@4bb1ed6b{/,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6de87c4c{/api,null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6355a0a0{/jobs/job/kill,null, > AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@580d6fd4{/stages/stage/kill, > null,AVAILABLE,@Spark} > 18/01/30 19:05:31 INFO server.ServerConnector: Started Spark@36b4f4c5 > {HTTP/1.1}{0.0.0.0:37939} > 18/01/30 19:05:31 INFO server.Server: Started @3740ms > 18/01/30 19:05:31 INFO util.Utils: Successfully started service 'SparkUI' > on port 37939. > 18/01/30 19:05:31 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and started > at http://192.168.1.184:37939 > 18/01/30 19:05:31 INFO cluster.YarnClusterScheduler: Created > YarnClusterScheduler > 18/01/30 19:05:31 INFO cluster.SchedulerExtensionServices: Starting Yarn > extension services with app application_1516010594436_0185 and attemptId > Some(appattempt_1516010594436_0185_000002) > 18/01/30 19:05:31 INFO util.Utils: Successfully started service > 'org.apache.spark.network.netty.NettyBlockTransferService' on port 36754. > 18/01/30 19:05:31 INFO netty.NettyBlockTransferService: Server created on > 192.168.1.184:36754 > 18/01/30 19:05:31 INFO storage.BlockManager: Using > org.apache.spark.storage.RandomBlockReplicationPolicy for block > replication policy > 18/01/30 19:05:31 INFO storage.BlockManagerMaster: Registering > BlockManager BlockManagerId(driver, 192.168.1.184, 36754, None) > 18/01/30 19:05:31 INFO storage.BlockManagerMasterEndpoint: Registering > block manager 192.168.1.184:36754 with 366.3 MB RAM, > BlockManagerId(driver, 192.168.1.184, 36754, None) > 18/01/30 19:05:31 INFO storage.BlockManagerMaster: Registered BlockManager > BlockManagerId(driver, 192.168.1.184, 36754, None) > 18/01/30 19:05:31 INFO storage.BlockManager: Initialized BlockManager: > BlockManagerId(driver, 192.168.1.184, 36754, None) > 18/01/30 19:05:31 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@57ef634b{/metrics/json,null, > AVAILABLE,@Spark} > 18/01/30 19:05:32 INFO scheduler.EventLoggingListener: Logging events to > hdfs:///kylin/spark-history/application_1516010594436_0185_2 > 18/01/30 19:05:32 INFO cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: > ApplicationMaster registered as NettyRpcEndpointRef(spark://Ya > [email protected]:42339) > 18/01/30 19:05:32 INFO yarn.ApplicationMaster: > ============================================================ > =================== > YARN executor launch context: > env: > CLASSPATH -> {{PWD}}<CPS>{{PWD}}/__spark_conf__<CPS>{{PWD}}/__spark_ > libs__/*<CPS>$HADOOP_CONF_DIR<CPS>$HADOOP_COMMON_HOME/share/ > hadoop/common/*<CPS>$HADOOP_COMMON_HOME/share/hadoop/ > common/lib/*<CPS>$HADOOP_HDFS_HOME/share/hadoop/hdfs/*<CPS>$ > HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*<CPS>$HADOOP_YARN_ > HOME/share/hadoop/yarn/*<CPS>$HADOOP_YARN_HOME/share/hadoop/ > yarn/lib/*<CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/*< > CPS>$HADOOP_MAPRED_HOME/share/hadoop/mapreduce/lib/* > SPARK_YARN_STAGING_DIR -> hdfs://192.168.1.171:9000/ > user/hadoop/.sparkStaging/application_1516010594436_0185 > SPARK_USER -> hadoop > SPARK_YARN_MODE -> true > > command: > {{JAVA_HOME}}/bin/java \ > -server \ > -Xmx1024m \ > '-Dhdp.version=current' \ > -Djava.io.tmpdir={{PWD}}/tmp \ > -Dspark.yarn.app.container.log.dir=<LOG_DIR> \ > -XX:OnOutOfMemoryError='kill %p' \ > org.apache.spark.executor.CoarseGrainedExecutorBackend \ > --driver-url \ > spark://[email protected]:42339 \ > --executor-id \ > <executorId> \ > --hostname \ > <hostname> \ > --cores \ > 2 \ > --app-id \ > application_1516010594436_0185 \ > --user-class-path \ > file:$PWD/__app__.jar \ > --user-class-path \ > file:$PWD/htrace-core-3.1.0-incubating.jar \ > --user-class-path \ > file:$PWD/metrics-core-2.2.0.jar \ > --user-class-path \ > file:$PWD/guava-12.0.1.jar \ > 1><LOG_DIR>/stdout \ > 2><LOG_DIR>/stderr > > resources: > htrace-core-3.1.0-incubating.jar -> resource { scheme: "hdfs" host: > "192.168.1.171" port: 9000 file: "/user/hadoop/.sparkStaging/ > application_1516010594436_0185/htrace-core-3.1.0-incubating.jar" } size: > 1475955 timestamp: 1517310297655 type: FILE visibility: PRIVATE > __app__.jar -> resource { scheme: "hdfs" host: "192.168.1.171" port: 9000 > file: > "/user/hadoop/.sparkStaging/application_1516010594436_0185/kylin-job-2.2.0.jar" > } size: 16695259 timestamp: 1517310297574 type: FILE visibility: PRIVATE > guava-12.0.1.jar -> resource { scheme: "hdfs" host: "192.168.1.171" port: > 9000 file: > "/user/hadoop/.sparkStaging/application_1516010594436_0185/guava-12.0.1.jar" > } size: 1795932 timestamp: 1517310297780 type: FILE visibility: PRIVATE > __spark_conf__ -> resource { scheme: "hdfs" host: "192.168.1.171" port: > 9000 file: > "/user/hadoop/.sparkStaging/application_1516010594436_0185/__spark_conf__.zip" > } size: 86203 timestamp: 1517310297898 type: ARCHIVE visibility: PRIVATE > metrics-core-2.2.0.jar -> resource { scheme: "hdfs" host: "192.168.1.171" > port: 9000 file: > "/user/hadoop/.sparkStaging/application_1516010594436_0185/metrics-core-2.2.0.jar" > } size: 82123 timestamp: 1517310297714 type: FILE visibility: PRIVATE > __spark_libs__ -> resource { scheme: "hdfs" host: "192.168.1.171" port: > 9000 file: "/kylin/spark/spark-libs.jar" } size: 200781646 timestamp: > 1517295215668 type: ARCHIVE visibility: PUBLIC > > ============================================================ > =================== > 18/01/30 19:05:32 INFO client.RMProxy: Connecting to ResourceManager at / > 192.168.1.171:8030 > 18/01/30 19:05:32 INFO yarn.YarnRMClient: Registering the ApplicationMaster > 18/01/30 19:05:32 INFO yarn.YarnAllocator: Will request 1 executor > container(s), each with 2 core(s) and 1408 MB memory (including 384 MB of > overhead) > 18/01/30 19:05:32 INFO yarn.YarnAllocator: Submitted 1 unlocalized > container requests. > 18/01/30 19:05:32 INFO yarn.ApplicationMaster: Started progress reporter > thread with (heartbeat : 3000, initial allocation : 200) intervals > 18/01/30 19:05:32 INFO impl.AMRMClientImpl: Received new token for : > sx-hadoop-dn02:32911 > 18/01/30 19:05:32 INFO yarn.YarnAllocator: Launching container > container_1516010594436_0185_02_000002 on host sx-hadoop-dn02 > 18/01/30 19:05:32 INFO yarn.YarnAllocator: Received 1 containers from > YARN, launching executors on 1 of them. > 18/01/30 19:05:32 INFO impl.ContainerManagementProtocolProxy: > yarn.client.max-cached-nodemanagers-proxies : 0 > 18/01/30 19:05:32 INFO impl.ContainerManagementProtocolProxy: Opening > proxy : sx-hadoop-dn02:32911 > 18/01/30 19:05:36 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: > Registered executor NettyRpcEndpointRef(null) (192.168.1.181:54402) with > ID 1 > 18/01/30 19:05:36 INFO cluster.YarnClusterSchedulerBackend: > SchedulerBackend is ready for scheduling beginning after reached > minRegisteredResourcesRatio: 0.8 > 18/01/30 19:05:36 INFO cluster.YarnClusterScheduler: > YarnClusterScheduler.postStartHook > done > 18/01/30 19:05:36 INFO storage.BlockManagerMasterEndpoint: Registering > block manager sx-hadoop-dn02:42882 with 366.3 MB RAM, BlockManagerId(1, > sx-hadoop-dn02, 42882, None) > 18/01/30 19:05:36 INFO common.AbstractHadoopJob: Ready to load KylinConfig > from uri: kylin_metadata@hdfs,path=hdfs://192.168.1.171:9000/kylin/ > kylin_metadata/metadata/c3468331-ab55-4368-b691-bf08e2b5fd2a > 18/01/30 19:05:36 INFO cube.CubeManager: Initializing CubeManager with > config kylin_metadata@hdfs,path=hdfs://192.168.1.171:9000/kylin/ > kylin_metadata/metadata/c3468331-ab55-4368-b691-bf08e2b5fd2a > 18/01/30 19:05:36 INFO persistence.ResourceStore: Using metadata url > kylin_metadata@hdfs,path=hdfs://192.168.1.171:9000/kylin/ > kylin_metadata/metadata/c3468331-ab55-4368-b691-bf08e2b5fd2a for resource > store > 18/01/30 19:05:36 INFO hdfs.HDFSResourceStore: hdfs meta path : hdfs:// > 192.168.1.171:9000/kylin/kylin_metadata/metadata/c3468331-ab55-4368-b691- > bf08e2b5fd2a > 18/01/30 19:05:36 INFO cube.CubeManager: Loading Cube from folder hdfs:// > 192.168.1.171:9000/kylin/kylin_metadata/metadata/c3468331-ab55-4368-b691- > bf08e2b5fd2a/cube > 18/01/30 19:05:37 INFO cube.CubeDescManager: Initializing CubeDescManager > with config kylin_metadata@hdfs,path=hdfs://192.168.1.171:9000/kylin/ > kylin_metadata/metadata/c3468331-ab55-4368-b691-bf08e2b5fd2a > 18/01/30 19:05:37 INFO cube.CubeDescManager: Reloading Cube Metadata from > folder hdfs://192.168.1.171:9000/kylin/kylin_metadata/metadata/ > c3468331-ab55-4368-b691-bf08e2b5fd2a/cube_desc > 18/01/30 19:05:37 INFO project.ProjectManager: Initializing > ProjectManager with metadata url kylin_metadata@hdfs,path=hdfs:// > 192.168.1.171:9000/kylin/kylin_metadata/metadata/c3468331-ab55-4368-b691- > bf08e2b5fd2a > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: Checking custom > measure types from kylin config > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering > COUNT_DISTINCT(hllc), class org.apache.kylin.measure.hllc. > HLLCMeasureType$Factory > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering > COUNT_DISTINCT(bitmap), class org.apache.kylin.measure. > bitmap.BitmapMeasureType$Factory > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering > TOP_N(topn), class org.apache.kylin.measure.topn.TopNMeasureType$Factory > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering RAW(raw), > class org.apache.kylin.measure.raw.RawMeasureType$Factory > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering > EXTENDED_COLUMN(extendedcolumn), class org.apache.kylin.measure. > extendedcolumn.ExtendedColumnMeasureType$Factory > 18/01/30 19:05:37 INFO measure.MeasureTypeFactory: registering > PERCENTILE(percentile), class org.apache.kylin.measure.percentile. > PercentileMeasureType$Factory > 18/01/30 19:05:37 INFO metadata.MetadataManager: Reloading data model at > /model_desc/kylin_sales_model.json > 18/01/30 19:05:37 INFO cube.CubeDescManager: Loaded 1 Cube(s) > 18/01/30 19:05:37 INFO cube.CubeManager: Reloaded cube > kylin_sales_cube_clone being CUBE[name=kylin_sales_cube_clone] having 2 > segments > 18/01/30 19:05:37 INFO cube.CubeManager: Loaded 1 cubes, fail on 0 cubes > 18/01/30 19:05:37 INFO spark.SparkCubingByLayer: RDD Output path: hdfs:// > 192.168.1.171:9000/kylin/kylin_metadata/kylin-5b2617ee-90ec-4c4b-86ff- > 699d0f14223e/kylin_sales_cube_clone/cuboid/ > 18/01/30 19:05:37 INFO spark.SparkCubingByLayer: All measure are normal > (agg on all cuboids) ? : true > 18/01/30 19:05:38 INFO internal.SharedState: Warehouse path is > 'file:/home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/application_ > 1516010594436_0185/container_1516010594436_0185_02_000001/ > spark-warehouse'. > 18/01/30 19:05:38 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@75fae875{/SQL,null,AVAILABLE,@Spark} > 18/01/30 19:05:38 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@3785cc60{/SQL/json,null,AVAILABLE,@Spark} > 18/01/30 19:05:38 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@6299f0a5{/SQL/execution,null, > AVAILABLE,@Spark} > 18/01/30 19:05:38 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@578d227{/SQL/execution/json, > null,AVAILABLE,@Spark} > 18/01/30 19:05:38 INFO handler.ContextHandler: Started > o.s.j.s.ServletContextHandler@29f2a18{/static/sql,null,AVAILABLE,@Spark} > 18/01/30 19:05:38 INFO hive.HiveUtils: Initializing > HiveMetastoreConnection version 1.2.1 using Spark classes. > 18/01/30 19:05:38 INFO metastore.HiveMetaStore: 0: Opening raw store with > implemenation class:org.apache.hadoop.hive.metastore.ObjectStore > 18/01/30 19:05:39 INFO metastore.ObjectStore: ObjectStore, initialize > called > 18/01/30 19:05:39 INFO DataNucleus.Persistence: Property > hive.metastore.integral.jdo.pushdown unknown - will be ignored > 18/01/30 19:05:39 INFO DataNucleus.Persistence: Property > datanucleus.cache.level2 unknown - will be ignored > 18/01/30 19:05:45 INFO metastore.ObjectStore: Setting MetaStore object pin > classes with hive.metastore.cache.pinobjtypes="Table, > StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" > 18/01/30 19:05:47 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 18/01/30 19:05:47 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 18/01/30 19:05:51 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as > "embedded-only" so does not have its own datastore table. > 18/01/30 19:05:51 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as > "embedded-only" so does not have its own datastore table. > 18/01/30 19:05:52 INFO metastore.MetaStoreDirectSql: Using direct SQL, > underlying DB is DERBY > 18/01/30 19:05:52 INFO metastore.ObjectStore: Initialized ObjectStore > 18/01/30 19:05:52 WARN metastore.ObjectStore: Version information not > found in metastore. hive.metastore.schema.verification is not enabled so > recording the schema version 1.2.0 > 18/01/30 19:05:52 WARN metastore.ObjectStore: Failed to get database > default, returning NoSuchObjectException > 18/01/30 19:05:53 INFO metastore.HiveMetaStore: Added admin role in > metastore > 18/01/30 19:05:53 INFO metastore.HiveMetaStore: Added public role in > metastore > 18/01/30 19:05:53 INFO metastore.HiveMetaStore: No user is added in admin > role, since config is empty > 18/01/30 19:05:53 INFO metastore.HiveMetaStore: 0: get_all_databases > 18/01/30 19:05:53 INFO HiveMetaStore.audit: ugi=hadoop > ip=unknown-ip-addr cmd=get_all_databases > 18/01/30 19:05:53 INFO metastore.HiveMetaStore: 0: get_functions: > db=default pat=* > 18/01/30 19:05:53 INFO HiveMetaStore.audit: ugi=hadoop > ip=unknown-ip-addr cmd=get_functions: db=default pat=* > 18/01/30 19:05:53 INFO DataNucleus.Datastore: The class > "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as > "embedded-only" so does not have its own datastore table. > 18/01/30 19:05:54 INFO session.SessionState: Created local directory: > /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/ > application_1516010594436_0185/container_1516010594436_ > 0185_02_000001/tmp/hadoop > 18/01/30 19:05:54 INFO session.SessionState: Created local directory: > /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/ > application_1516010594436_0185/container_1516010594436_ > 0185_02_000001/tmp/4990ced4-7b5c-46d6-a0f6-9ecfdcbc22c5_resources > 18/01/30 19:05:54 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hadoop/4990ced4-7b5c-46d6-a0f6-9ecfdcbc22c5 > 18/01/30 19:05:54 INFO session.SessionState: Created local directory: > /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/ > application_1516010594436_0185/container_1516010594436_ > 0185_02_000001/tmp/hadoop/4990ced4-7b5c-46d6-a0f6-9ecfdcbc22c5 > 18/01/30 19:05:54 INFO session.SessionState: Created HDFS directory: > /tmp/hive/hadoop/4990ced4-7b5c-46d6-a0f6-9ecfdcbc22c5/_tmp_space.db > 18/01/30 19:05:54 INFO client.HiveClientImpl: Warehouse location for Hive > client (version 1.2.1) is file:/home/hadoop/tmp/nm- > local-dir/usercache/hadoop/appcache/application_ > 1516010594436_0185/container_1516010594436_0185_02_000001/spark-warehouse > 18/01/30 19:05:54 INFO metastore.HiveMetaStore: 0: get_database: default > 18/01/30 19:05:54 INFO HiveMetaStore.audit: ugi=hadoop > ip=unknown-ip-addr cmd=get_database: default > 18/01/30 19:05:54 INFO metastore.HiveMetaStore: 0: get_database: > global_temp > 18/01/30 19:05:54 INFO HiveMetaStore.audit: ugi=hadoop > ip=unknown-ip-addr cmd=get_database: global_temp > 18/01/30 19:05:54 WARN metastore.ObjectStore: Failed to get database > global_temp, returning NoSuchObjectException > 18/01/30 19:05:54 INFO execution.SparkSqlParser: Parsing command: > default.kylin_intermediate_kylin_sales_cube_clone_c3468331_ab55_4368_b691_ > bf08e2b5fd2a > 18/01/30 19:05:55 INFO metastore.HiveMetaStore: 0: get_table : db=default > tbl=kylin_intermediate_kylin_sales_cube_clone_c3468331_ > ab55_4368_b691_bf08e2b5fd2a > 18/01/30 19:05:55 INFO HiveMetaStore.audit: ugi=hadoop > ip=unknown-ip-addr cmd=get_table : db=default > tbl=kylin_intermediate_kylin_sales_cube_clone_c3468331_ > ab55_4368_b691_bf08e2b5fd2a > 18/01/30 19:05:55 ERROR yarn.ApplicationMaster: User class threw > exception: java.lang.RuntimeException: error execute > org.apache.kylin.engine.spark.SparkCubingByLayer > java.lang.RuntimeException: error execute org.apache.kylin.engine.spark. > SparkCubingByLayer > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:42) > at org.apache.kylin.common.util.SparkEntry.main(SparkEntry. > java:44) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:62) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run( > ApplicationMaster.scala:636) > Caused by: org.apache.spark.sql.catalyst.analysis.NoSuchTableException: > Table or view > 'kylin_intermediate_kylin_sales_cube_clone_c3468331_ab55_4368_b691_bf08e2b5fd2a' > not found in database 'default'; > at org.apache.spark.sql.hive.client.HiveClient$$anonfun$ > getTable$1.apply(HiveClient.scala:74) > at org.apache.spark.sql.hive.client.HiveClient$$anonfun$ > getTable$1.apply(HiveClient.scala:74) > at scala.Option.getOrElse(Option.scala:121) > at org.apache.spark.sql.hive.client.HiveClient$class. > getTable(HiveClient.scala:74) > at org.apache.spark.sql.hive.client.HiveClientImpl. > getTable(HiveClientImpl.scala:78) > at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$ > org$apache$spark$sql$hive$HiveExternalCatalog$$getRawTable$1.apply( > HiveExternalCatalog.scala:118) > at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$ > org$apache$spark$sql$hive$HiveExternalCatalog$$getRawTable$1.apply( > HiveExternalCatalog.scala:118) > at org.apache.spark.sql.hive.HiveExternalCatalog.withClient( > HiveExternalCatalog.scala:97) > at org.apache.spark.sql.hive.HiveExternalCatalog.org$ > apache$spark$sql$hive$HiveExternalCatalog$$getRawTable( > HiveExternalCatalog.scala:117) > at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$ > getTable$1.apply(HiveExternalCatalog.scala:628) > at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$ > getTable$1.apply(HiveExternalCatalog.scala:628) > at org.apache.spark.sql.hive.HiveExternalCatalog.withClient( > HiveExternalCatalog.scala:97) > at org.apache.spark.sql.hive.HiveExternalCatalog.getTable( > HiveExternalCatalog.scala:627) > at org.apache.spark.sql.hive.HiveMetastoreCatalog.lookupRelation( > HiveMetastoreCatalog.scala:124) > at org.apache.spark.sql.hive.HiveSessionCatalog.lookupRelation( > HiveSessionCatalog.scala:70) > at org.apache.spark.sql.SparkSession.table(SparkSession.scala:586) > at org.apache.spark.sql.SparkSession.table(SparkSession.scala:582) > at org.apache.spark.sql.SQLContext.table(SQLContext.scala:708) > at org.apache.kylin.engine.spark.SparkCubingByLayer.execute( > SparkCubingByLayer.java:167) > at org.apache.kylin.common.util.AbstractApplication.execute( > AbstractApplication.java:37) > ... 6 more > 18/01/30 19:05:55 INFO yarn.ApplicationMaster: Final app status: FAILED, > exitCode: 15, (reason: User class threw exception: > java.lang.RuntimeException: error execute org.apache.kylin.engine.spark. > SparkCubingByLayer) > 18/01/30 19:05:55 INFO spark.SparkContext: Invoking stop() from shutdown > hook > 18/01/30 19:05:55 INFO server.ServerConnector: Stopped Spark@36b4f4c5 > {HTTP/1.1}{0.0.0.0:0} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@580d6fd4{/stages/stage/kill, > null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@6355a0a0{/jobs/job/kill,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@6de87c4c{/api,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@4bb1ed6b{/,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@d66c0ae{/static,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@3de67ab3{/executors/threadDump/json,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@14c64726{/executors/ > threadDump,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@1afb9d54{/executors/json,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@7bd97d3c{/executors,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@6af87e2b{/environment/json, > null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@4215247{/environment,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@53e654bf{/storage/rdd/json, > null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@2d479ee6{/storage/rdd,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@a2c9363{/storage/json,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@71f4dc98{/storage,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@1e1c0871{/stages/pool/json, > null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@41ccc000{/stages/pool,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@6d72da27{/stages/stage/json, > null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@4b84807a{/stages/stage,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@44336797{/stages/json,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@7f3a751c{/stages,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@6dc9732e{/jobs/job/json,null, > UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@201b73c{/jobs/job,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@5a8c27d3{/jobs/json,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO handler.ContextHandler: Stopped > o.s.j.s.ServletContextHandler@1727271{/jobs,null,UNAVAILABLE,@Spark} > 18/01/30 19:05:55 INFO ui.SparkUI: Stopped Spark web UI at > http://192.168.1.184:37939 > 18/01/30 19:05:55 INFO yarn.YarnAllocator: Driver requested a total number > of 0 executor(s). > 18/01/30 19:05:55 INFO cluster.YarnClusterSchedulerBackend: Shutting down > all executors > 18/01/30 19:05:55 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: > Asking each executor to shut down > 18/01/30 19:05:55 INFO cluster.SchedulerExtensionServices: Stopping > SchedulerExtensionServices > (serviceOption=None, > services=List(), > started=false) > 18/01/30 19:05:55 INFO spark.MapOutputTrackerMasterEndpoint: > MapOutputTrackerMasterEndpoint stopped! > 18/01/30 19:05:55 INFO memory.MemoryStore: MemoryStore cleared > 18/01/30 19:05:55 INFO storage.BlockManager: BlockManager stopped > 18/01/30 19:05:55 INFO storage.BlockManagerMaster: BlockManagerMaster > stopped > 18/01/30 19:05:55 INFO scheduler.OutputCommitCoordinator$ > OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! > 18/01/30 19:05:55 INFO spark.SparkContext: Successfully stopped > SparkContext > 18/01/30 19:05:55 INFO yarn.ApplicationMaster: Unregistering > ApplicationMaster with FAILED (diag message: User class threw exception: > java.lang.RuntimeException: error execute org.apache.kylin.engine.spark. > SparkCubingByLayer) > 18/01/30 19:05:55 INFO impl.AMRMClientImpl: Waiting for application to be > successfully unregistered. > 18/01/30 19:05:55 INFO yarn.ApplicationMaster: Deleting staging directory > hdfs://192.168.1.171:9000/user/hadoop/.sparkStaging/ > application_1516010594436_0185 > 18/01/30 19:05:55 INFO util.ShutdownHookManager: Shutdown hook called > 18/01/30 19:05:55 INFO util.ShutdownHookManager: Deleting directory > /home/hadoop/tmp/nm-local-dir/usercache/hadoop/appcache/ > application_1516010594436_0185/spark-502e0062-d08e-47a0-9663-4b284f141d3f > > > -- Best regards, Shaofeng Shi 史少锋
