AfterYesterday opened a new issue #3654:
URL: https://github.com/apache/incubator-dolphinscheduler/issues/3654


   [INFO] 2020-09-03 10:57:31.411  - [taskAppId=TASK-3-3-7]:[121] -  -> 
Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" 
with specified deploy mode instead.
        20/09/03 10:57:31 INFO spark.SparkContext: Running Spark version 
2.4.0.cloudera2
   [INFO] 2020-09-03 10:57:33.234  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:31 INFO spark.SparkContext: Submitted application: sparkSqlTest
        20/09/03 10:57:31 INFO spark.SecurityManager: Changing view acls to: 
hdfs
        20/09/03 10:57:31 INFO spark.SecurityManager: Changing modify acls to: 
hdfs
        20/09/03 10:57:31 INFO spark.SecurityManager: Changing view acls groups 
to: 
        20/09/03 10:57:31 INFO spark.SecurityManager: Changing modify acls 
groups to: 
        20/09/03 10:57:31 INFO spark.SecurityManager: SecurityManager: 
authentication disabled; ui acls disabled; users  with view permissions: 
Set(hdfs); groups with view permissions: Set(); users  with modify permissions: 
Set(hdfs); groups with modify permissions: Set()
        20/09/03 10:57:31 INFO util.Utils: max retries is 16
        20/09/03 10:57:31 INFO util.Utils: Successfully started service 
'sparkDriver' on port 35977.
        20/09/03 10:57:31 INFO spark.SparkEnv: Registering MapOutputTracker
        20/09/03 10:57:31 INFO spark.SparkEnv: Registering BlockManagerMaster
        20/09/03 10:57:31 INFO storage.BlockManagerMasterEndpoint: Using 
org.apache.spark.storage.DefaultTopologyMapper for getting topology information
        20/09/03 10:57:31 INFO storage.BlockManagerMasterEndpoint: 
BlockManagerMasterEndpoint up
        20/09/03 10:57:31 INFO storage.DiskBlockManager: Created local 
directory at /tmp/blockmgr-c87ef2c5-a62b-4e48-a849-d70660e4d66e
        20/09/03 10:57:31 INFO memory.MemoryStore: MemoryStore started with 
capacity 366.3 MB
        20/09/03 10:57:31 INFO spark.SparkEnv: Registering 
OutputCommitCoordinator
        20/09/03 10:57:31 INFO util.log: Logging initialized @2045ms
        20/09/03 10:57:32 INFO server.Server: jetty-9.3.z-SNAPSHOT, build 
timestamp: unknown, git hash: unknown
        20/09/03 10:57:32 INFO server.Server: Started @2119ms
        20/09/03 10:57:32 INFO util.Utils: max retries is 16
        20/09/03 10:57:32 INFO server.AbstractConnector: Started 
ServerConnector@7f4d9395{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
        20/09/03 10:57:32 INFO util.Utils: Successfully started service 
'SparkUI' on port 4040.
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@3003697{/jobs,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@2aa27288{/jobs/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@7f34a967{/jobs/job,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@240139e1{/jobs/job/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@7ea4d397{/stages,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@49298ce7{/stages/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@253c1256{/stages/stage,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@55f45b92{/stages/stage/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@109f5dd8{/stages/pool,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@67fe380b{/stages/pool/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@4a325eb9{/storage,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@3dedb4a6{/storage/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@57f64f5e{/storage/rdd,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@415e0bcb{/storage/rdd/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@194152cf{/environment,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@49d98dc5{/environment/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@2c30b71f{/executors,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@1d81e101{/executors/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@ec50f54{/executors/threadDump,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@bf71cec{/executors/threadDump/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@22d6cac2{/static,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@6d868997{/,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@2c383e33{/api,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@4287d447{/jobs/job/kill,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@3af37506{/stages/stage/kill,null,AVAILABLE,@Spark}
        20/09/03 10:57:32 INFO ui.SparkUI: Bound SparkUI to 0.0.0.0, and 
started at http://dsjzt-hadoop-2:4040
        20/09/03 10:57:32 INFO spark.SparkContext: Added JAR 
file:/sparkjob/scalatest-1.0-SNAPSHOT.jar at 
spark://dsjzt-hadoop-2:35977/jars/scalatest-1.0-SNAPSHOT.jar with timestamp 
1599101852188
        20/09/03 10:57:32 INFO util.Utils: Using initial executors = 2, max of 
spark.dynamicAllocation.initialExecutors, spark.dynamicAllocation.minExecutors 
and spark.executor.instances
        20/09/03 10:57:33 INFO client.RMProxy: Connecting to ResourceManager at 
dsjzt-hadoop-1/10.7.11.9:8032
   [INFO] 2020-09-03 10:57:35.226  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:33 INFO yarn.Client: Requesting a new application from cluster 
with 3 NodeManagers
        20/09/03 10:57:33 INFO yarn.Client: Verifying our application has not 
requested more than the maximum memory capability of the cluster (48119 MB per 
container)
        20/09/03 10:57:33 INFO yarn.Client: Will allocate AM container, with 
896 MB memory including 384 MB overhead
        20/09/03 10:57:33 INFO yarn.Client: Setting up container launch context 
for our AM
        20/09/03 10:57:33 INFO yarn.Client: Setting up the launch environment 
for our AM container
        20/09/03 10:57:33 INFO yarn.Client: Preparing resources for our AM 
container
        20/09/03 10:57:33 INFO yarn.Client: Uploading resource 
file:/tmp/spark-60cfad97-9e20-460f-b947-cce3b9092bb2/__spark_conf__1905511865034791586.zip
 -> 
hdfs://dsjzt-db:8020/user/hdfs/.sparkStaging/application_1598319282181_0237/__spark_conf__.zip
        20/09/03 10:57:34 INFO spark.SecurityManager: Changing view acls to: 
hdfs
        20/09/03 10:57:34 INFO spark.SecurityManager: Changing modify acls to: 
hdfs
        20/09/03 10:57:34 INFO spark.SecurityManager: Changing view acls groups 
to: 
        20/09/03 10:57:34 INFO spark.SecurityManager: Changing modify acls 
groups to: 
        20/09/03 10:57:34 INFO spark.SecurityManager: SecurityManager: 
authentication disabled; ui acls disabled; users  with view permissions: 
Set(hdfs); groups with view permissions: Set(); users  with modify permissions: 
Set(hdfs); groups with modify permissions: Set()
        20/09/03 10:57:35 INFO yarn.Client: Submitting application 
application_1598319282181_0237 to ResourceManager
   [INFO] 2020-09-03 10:57:36.272  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:35 INFO impl.YarnClientImpl: Submitted application 
application_1598319282181_0237
        20/09/03 10:57:35 INFO cluster.SchedulerExtensionServices: Starting 
Yarn extension services with app application_1598319282181_0237 and attemptId 
None
        20/09/03 10:57:35 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
        20/09/03 10:57:36 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
   [INFO] 2020-09-03 10:57:37.278  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:36 INFO yarn.Client: Application report for 
application_1598319282181_0237 (state: ACCEPTED)
        20/09/03 10:57:36 INFO yarn.Client: 
                 client token: N/A
                 diagnostics: N/A
                 ApplicationMaster host: N/A
                 ApplicationMaster RPC port: -1
                 queue: root.users.hdfs
                 start time: 1599101850311
                 final status: UNDEFINED
                 tracking URL: 
http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237/
                 user: hdfs
        20/09/03 10:57:37 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
        20/09/03 10:57:37 INFO yarn.Client: Application report for 
application_1598319282181_0237 (state: ACCEPTED)
   [INFO] 2020-09-03 10:57:38.280  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:38 INFO yarn.SparkRackResolver: Got an error when resolving 
hostNames. Falling back to /default-rack for all
        20/09/03 10:57:38 INFO yarn.Client: Application report for 
application_1598319282181_0237 (state: ACCEPTED)
   [INFO] 2020-09-03 10:57:39.282  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:38 INFO cluster.YarnClientSchedulerBackend: Add WebUI Filter. 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> 
dsjzt-hadoop-1, PROXY_URI_BASES -> 
http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237), 
/proxy/application_1598319282181_0237
        20/09/03 10:57:38 INFO ui.JettyUtils: Adding filter 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /jobs, 
/jobs/json, /jobs/job, /jobs/job/json, /stages, /stages/json, /stages/stage, 
/stages/stage/json, /stages/pool, /stages/pool/json, /storage, /storage/json, 
/storage/rdd, /storage/rdd/json, /environment, /environment/json, /executors, 
/executors/json, /executors/threadDump, /executors/threadDump/json, /static, /, 
/api, /jobs/job/kill, /stages/stage/kill.
        20/09/03 10:57:39 INFO 
cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: ApplicationMaster 
registered as NettyRpcEndpointRef(spark-client://YarnAM)
        20/09/03 10:57:39 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
        20/09/03 10:57:39 INFO yarn.Client: Application report for 
application_1598319282181_0237 (state: RUNNING)
   [INFO] 2020-09-03 10:57:41.272  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:39 INFO yarn.Client: 
                 client token: N/A
                 diagnostics: N/A
                 ApplicationMaster host: 10.7.11.9
                 ApplicationMaster RPC port: -1
                 queue: root.users.hdfs
                 start time: 1599101850311
                 final status: UNDEFINED
                 tracking URL: 
http://dsjzt-hadoop-1:8088/proxy/application_1598319282181_0237/
                 user: hdfs
        20/09/03 10:57:39 INFO cluster.YarnClientSchedulerBackend: Application 
application_1598319282181_0237 has started running.
        20/09/03 10:57:39 INFO util.Utils: max retries is 16
        20/09/03 10:57:39 INFO util.Utils: Successfully started service 
'org.apache.spark.network.netty.NettyBlockTransferService' on port 40754.
        20/09/03 10:57:39 INFO netty.NettyBlockTransferService: Server created 
on dsjzt-hadoop-2:40754
        20/09/03 10:57:39 INFO storage.BlockManager: Using 
org.apache.spark.storage.RandomBlockReplicationPolicy for block replication 
policy
        20/09/03 10:57:39 INFO storage.BlockManagerMaster: Registering 
BlockManager BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
        20/09/03 10:57:39 INFO storage.BlockManagerMasterEndpoint: Registering 
block manager dsjzt-hadoop-2:40754 with 366.3 MB RAM, BlockManagerId(driver, 
dsjzt-hadoop-2, 40754, None)
        20/09/03 10:57:39 INFO storage.BlockManagerMaster: Registered 
BlockManager BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
        20/09/03 10:57:39 INFO storage.BlockManager: external shuffle service 
port = 7337
        20/09/03 10:57:39 INFO storage.BlockManager: Initialized BlockManager: 
BlockManagerId(driver, dsjzt-hadoop-2, 40754, None)
        20/09/03 10:57:39 INFO ui.JettyUtils: Adding filter 
org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /metrics/json.
        20/09/03 10:57:39 INFO handler.ContextHandler: Started 
o.s.j.s.ServletContextHandler@191ec193{/metrics/json,null,AVAILABLE,@Spark}
        20/09/03 10:57:39 INFO scheduler.EventLoggingListener: Logging events 
to 
hdfs://dsjzt-db:8020/user/spark/spark2ApplicationHistory/application_1598319282181_0237
        20/09/03 10:57:39 INFO util.Utils: Using initial executors = 2, max of 
spark.dynamicAllocation.initialExecutors, spark.dynamicAllocation.minExecutors 
and spark.executor.instances
        20/09/03 10:57:39 WARN lineage.LineageWriter: Lineage directory 
/var/log/spark2/lineage doesn't exist or is not writable. Lineage for this 
application will be disabled.
        20/09/03 10:57:39 INFO util.Utils: Extension 
com.cloudera.spark.lineage.NavigatorAppListener not being initialized.
        20/09/03 10:57:40 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
        20/09/03 10:57:41 INFO yarn.SparkRackResolver: Got an error when 
resolving hostNames. Falling back to /default-rack for all
   [INFO] 2020-09-03 10:57:42.491  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:42 INFO yarn.SparkRackResolver: Got an error when resolving 
hostNames. Falling back to /default-rack for all
        20/09/03 10:57:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: 
Registered executor NettyRpcEndpointRef(spark-client://Executor) 
(10.7.11.9:34350) with ID 1
   [INFO] 2020-09-03 10:57:45.199  - [taskAppId=TASK-3-3-7]:[121] -  -> 
20/09/03 10:57:42 INFO spark.ExecutorAllocationManager: New executor 1 has 
registered (new total is 1)
        20/09/03 10:57:42 INFO cluster.YarnSchedulerBackend$YarnDriverEndpoint: 
Registered executor NettyRpcEndpointRef(spark-client://Executor) 
(10.7.11.9:34352) with ID 2
        20/09/03 10:57:42 INFO spark.ExecutorAllocationManager: New executor 2 
has registered (new total is 2)
        20/09/03 10:57:42 INFO storage.BlockManagerMasterEndpoint: Registering 
block manager dsjzt-hadoop-1:37298 with 5.2 GB RAM, BlockManagerId(1, 
dsjzt-hadoop-1, 37298, None)
        20/09/03 10:57:42 INFO cluster.YarnClientSchedulerBackend: 
SchedulerBackend is ready for scheduling beginning after reached 
minRegisteredResourcesRatio: 0.8
        20/09/03 10:57:43 WARN lineage.LineageWriter: Lineage directory 
/var/log/spark2/lineage doesn't exist or is not writable. Lineage for this 
application will be disabled.
        Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] 
java.io.FileNotFoundException: derby.log (Permission denied)
   [INFO] 2020-09-03 10:57:45.330  - [taskAppId=TASK-3-3-7]:[121] -  -> Thu Sep 
03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
        ERROR XBM0H: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown 
Source)
                at java.security.AccessController.doPrivileged(Native Method)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
   [INFO] 2020-09-03 10:57:45.331  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
                at 
scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
                at scala.collection.immutable.List.foldLeft(List.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
                at scala.collection.immutable.List.foreach(List.scala:392)
   [INFO] 2020-09-03 10:57:45.335  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
                at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
                at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
                at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
                at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
                at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
                at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
                at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
                at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
                at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
                at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
                at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
                at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
        Cleanup action completed
        Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
        java.sql.SQLException: Failed to create database 'metastore_db', see 
the next exception for details.
                at 
org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
                at 
org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
                at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
   [INFO] 2020-09-03 10:57:45.336  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
   [INFO] 2020-09-03 10:57:45.336  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
                at 
scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
                at scala.collection.immutable.List.foldLeft(List.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
                at scala.collection.immutable.List.foreach(List.scala:392)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
                at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
                at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
                at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
                at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
                at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
                at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
                at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
                at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
                at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
                at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
                at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
                at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
        Caused by: ERROR XJ041: Failed to create database 'metastore_db', see 
the next exception for details.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.impl.jdbc.SQLExceptionFactory.wrapArgsForTransportAcrossDRDA(Unknown
 Source)
                ... 146 more
        Caused by: ERROR XBM0H: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown 
Source)
                at java.security.AccessController.doPrivileged(Native Method)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown 
Source)
                ... 143 more
        ============= begin nested exception, level (1) ===========
        java.sql.SQLException: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
   [INFO] 2020-09-03 10:57:45.337  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.Util.generateCsSQLException(Unknown Source)
                at 
org.apache.derby.impl.jdbc.TransactionResourceImpl.wrapInSQLException(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.TransactionResourceImpl.handleException(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.handleException(Unknown Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
   [INFO] 2020-09-03 10:57:45.337  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
                at 
scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
                at scala.collection.immutable.List.foldLeft(List.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
                at scala.collection.immutable.List.foreach(List.scala:392)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
                at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
                at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
                at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
                at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
                at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
                at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
                at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
                at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
                at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
                at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
                at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
                at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
        Caused by: ERROR XBM0H: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown 
Source)
                at java.security.AccessController.doPrivileged(Native Method)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown 
Source)
                ... 143 more
        ============= end nested exception, level (1) ===========
        ============= begin nested exception, level (2) ===========
        ERROR XBM0H: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown 
Source)
                at java.security.AccessController.doPrivileged(Native Method)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:349)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
   [INFO] 2020-09-03 10:57:45.338  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
                at 
scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
                at scala.collection.immutable.List.foldLeft(List.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
                at scala.collection.immutable.List.foreach(List.scala:392)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
                at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
                at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
                at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
                at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
                at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
                at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
                at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
                at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
                at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
                at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
                at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
                at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
        ============= end nested exception, level (2) ===========
        Cleanup action completed
        Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
        ERROR XBM0H: Directory 
/tmp/dolphinscheduler/exec/process/1/3/3/7/metastore_db cannot be created.
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
                at 
org.apache.derby.iapi.error.StandardException.newException(Unknown Source)
   [INFO] 2020-09-03 10:57:45.339  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at org.apache.derby.impl.services.monitor.StorageFactoryService$10.run(Unknown 
Source)
                at java.security.AccessController.doPrivileged(Native Method)
                at 
org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown Source)
                at 
org.apache.derby.impl.services.monitor.BaseMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.impl.services.monitor.FileMonitor.createPersistentService(Unknown
 Source)
                at 
org.apache.derby.iapi.services.monitor.Monitor.createPersistentService(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
   [INFO] 2020-09-03 10:57:45.339  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager$lzycompute(SharedState.scala:141)
                at 
org.apache.spark.sql.internal.SharedState.globalTempViewManager(SharedState.scala:136)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.hive.HiveSessionStateBuilder$$anonfun$2.apply(HiveSessionStateBuilder.scala:55)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager$lzycompute(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.globalTempViewManager(SessionCatalog.scala:91)
                at 
org.apache.spark.sql.catalyst.catalog.SessionCatalog.isTemporaryTable(SessionCatalog.scala:736)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.isRunningDirectlyOnFiles(Analyzer.scala:749)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:683)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:715)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$apply$1.apply(AnalysisHelper.scala:90)
                at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:89)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1$$anonfun$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:326)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187)
                at 
org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:324)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:87)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$$anonfun$resolveOperatorsUp$1.apply(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:194)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$class.resolveOperatorsUp(AnalysisHelper.scala:86)
                at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperatorsUp(LogicalPlan.scala:29)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:708)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:654)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:87)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:84)
                at 
scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124)
                at scala.collection.immutable.List.foldLeft(List.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:84)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:76)
                at scala.collection.immutable.List.foreach(List.scala:392)
                at 
org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:76)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.org$apache$spark$sql$catalyst$analysis$Analyzer$$executeSameContext(Analyzer.scala:127)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.execute(Analyzer.scala:121)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:106)
   [INFO] 2020-09-03 10:57:45.340  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at 
org.apache.spark.sql.catalyst.analysis.Analyzer$$anonfun$executeAndCheck$1.apply(Analyzer.scala:105)
                at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
                at 
org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:105)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
                at 
org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
                at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
                at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:78)
                at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
                at dsjzt.sparkSqlTest$.main(sparkSqlTest.scala:14)
                at dsjzt.sparkSqlTest.main(sparkSqlTest.scala)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
                at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
                at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
                at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
                at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
                at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
                at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
                at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
        Cleanup action completed
        Thu Sep 03 10:57:45 CST 2020 Thread[main,5,main] Cleanup action starting
        java.sql.SQLException: Failed to create database 'metastore_db', see 
the next exception for details.
                at 
org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
                at 
org.apache.derby.impl.jdbc.SQLExceptionFactory.getSQLException(Unknown Source)
                at org.apache.derby.impl.jdbc.Util.seeNextException(Unknown 
Source)
                at 
org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown Source)
                at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown 
Source)
                at 
org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
                at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown 
Source)
                at java.sql.DriverManager.getConnection(DriverManager.java:664)
                at java.sql.DriverManager.getConnection(DriverManager.java:208)
                at 
com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(BoneCP.java:361)
                at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
                at 
com.jolbox.bonecp.BoneCPDataSource.getConnection(BoneCPDataSource.java:120)
                at 
org.datanucleus.store.rdbms.ConnectionFactoryImpl$ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
                at 
org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(RDBMSStoreManager.java:298)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.datanucleus.plugin.NonManagedPluginRegistry.createExecutableExtension(NonManagedPluginRegistry.java:631)
                at 
org.datanucleus.plugin.PluginManager.createExecutableExtension(PluginManager.java:301)
                at 
org.datanucleus.NucleusContext.createStoreManagerForProperties(NucleusContext.java:1187)
                at 
org.datanucleus.NucleusContext.initialise(NucleusContext.java:356)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.freezeConfiguration(JDOPersistenceManagerFactory.java:775)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
                at 
org.datanucleus.api.jdo.JDOPersistenceManagerFactory.getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
                at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
                at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                at java.lang.reflect.Method.invoke(Method.java:483)
                at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
                at java.security.AccessController.doPrivileged(Native Method)
                at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
                at 
javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1166)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
                at 
javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:419)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:448)
   [INFO] 2020-09-03 10:57:45.340  - [taskAppId=TASK-3-3-7]:[121] -  ->         
at org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:343)
                at 
org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:299)
                at 
org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
                at 
org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:60)
                at 
org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:69)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:685)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:663)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:712)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:511)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:78)
                at 
org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:84)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:6517)
                at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:207)
                at 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
                at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
                at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
                at 
java.lang.reflect.Constructor.newInstance(Constructor.java:408)
                at 
org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1660)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:67)
                at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:82)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.createMetaStoreClient(Hive.java:3411)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getMSC(Hive.java:3430)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.getAllFunctions(Hive.java:3655)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.reloadFunctions(Hive.java:231)
                at 
org.apache.hadoop.hive.ql.metadata.Hive.registerAllFunctionsOnce(Hive.java:215)
                at org.apache.hadoop.hive.ql.metadata.Hive.<init>(Hive.java:338)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:299)
                at org.apache.hadoop.hive.ql.metadata.Hive.get(Hive.java:274)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.org$apache$spark$sql$hive$client$HiveClientImpl$$client(HiveClientImpl.scala:246)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl$$anonfun$withHiveState$1.apply(HiveClientImpl.scala:268)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.liftedTree1$1(HiveClientImpl.scala:213)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.retryLocked(HiveClientImpl.scala:212)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.withHiveState(HiveClientImpl.scala:258)
                at 
org.apache.spark.sql.hive.client.HiveClientImpl.databaseExists(HiveClientImpl.scala:348)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:217)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:99)
                at 
org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:216)
                at 
org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
   
=================================================================================
   1:/tmp/dolphinscheduler已经给dolphinscheduler用户赋权过了
   2:mysql的jdbc包也放在对应的lib下了
   3:shell命令、hadoop命令、hive -e命令都是没问题的,可以跑
   请问还有什么其他的地方会影响到这个吗?


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to