你好,我看log里连接的是10000端口,这个是HS2的端口吧?Flink的HiveCatalog需要连接的是HMS,可以启动一个HMS再试试哈。
On Tue, Oct 27, 2020 at 9:57 AM RS <[email protected]> wrote: > Hi, 请教下 > 我尝试使用sql-client连接hive, hive正常, 使用beeline -u jdbc:hive2://x.x.x.x:10000 > 可以正常连接 > > > sql-client-defaults.yaml配置内容: > tables: [] > functions: [] > catalogs: > - name: myhive > type: hive > hive-conf-dir: /home/hive/flink-1.11.1/conf > default-database: default > execution: > planner: blink > type: streaming > time-characteristic: event-time > periodic-watermarks-interval: 200 > result-mode: table > max-table-result-rows: 1000000 > parallelism: 1 > max-parallelism: 128 > min-idle-state-retention: 0 > max-idle-state-retention: 0 > restart-strategy: > type: fallback > deployment: > response-timeout: 5000 > gateway-address: "" > gateway-port: 0 > > > 然后启动sql-client报错 > $./bin/sql-client.sh embedded > > > 最后的报错信息: > Exception in thread "main" > org.apache.flink.table.client.SqlClientException: Unexpected exception. > This is a bug. Please consider filing an issue. > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:213) > Caused by: org.apache.flink.table.client.gateway.SqlExecutionException: > Could not create execution context. > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:870) > at > org.apache.flink.table.client.gateway.local.LocalExecutor.openSession(LocalExecutor.java:227) > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:108) > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201) > Caused by: org.apache.flink.table.catalog.exceptions.CatalogException: > Failed to determine whether database default exists or not > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:335) > at > org.apache.flink.table.catalog.hive.HiveCatalog.open(HiveCatalog.java:227) > at > org.apache.flink.table.catalog.CatalogManager.registerCatalog(CatalogManager.java:191) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.registerCatalog(TableEnvironmentImpl.java:337) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$null$5(ExecutionContext.java:627) > at java.util.HashMap.forEach(HashMap.java:1289) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$initializeCatalogs$6(ExecutionContext.java:625) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:264) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:624) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:523) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:183) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:136) > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:859) > ... 3 more > Caused by: org.apache.thrift.transport.TTransportException > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > at > org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429) > at > org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318) > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219) > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:1135) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:1122) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1511) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1506) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:208) > at com.sun.proxy.$Proxy28.getDatabase(Unknown Source) > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.getDatabase(HiveMetastoreClientWrapper.java:107) > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:330) > ... 15 more > > > > > 附录完整错误信息: > Searching for > '/home/hive/flink-1.11.1/conf/sql-client-defaults.yaml'...found. > Reading default environment from: > file:/home/hive/flink-1.11.1/conf/sql-client-defaults.yaml > No session environment specified. > 2020-10-27 09:48:14,533 INFO org.apache.hadoop.hive.conf.HiveConf > [] - Found configuration file > file:/home/hive/flink-1.11.1/conf/hive-site.xml > 2020-10-27 09:48:15,144 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Trying to > connect to metastore with URI thrift://x.x.x.x:10000 > 2020-10-27 09:48:15,168 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Opened a > connection to metastore, current connections: 1 > 2020-10-27 09:48:15,240 WARN > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - set_ugi() > not successful, Likely cause: new client talking to old server. Continuing > without it. > org.apache.thrift.transport.TTransportException: null > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:380) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:230) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4787) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4773) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:534) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:224) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > ~[?:1.8.0_251] > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > ~[?:1.8.0_251] > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > ~[?:1.8.0_251] > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > ~[?:1.8.0_251] > at > org.apache.hadoop.hive.metastore.utils.JavaUtils.newInstance(JavaUtils.java:84) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:95) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:148) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[?:1.8.0_251] > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:1.8.0_251] > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:1.8.0_251] > at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_251] > at > org.apache.flink.table.catalog.hive.client.HiveShimV310.getHiveMetastoreClient(HiveShimV310.java:103) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.createMetastoreClient(HiveMetastoreClientWrapper.java:240) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.<init>(HiveMetastoreClientWrapper.java:71) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientFactory.create(HiveMetastoreClientFactory.java:35) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.HiveCatalog.open(HiveCatalog.java:223) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.CatalogManager.registerCatalog(CatalogManager.java:191) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.registerCatalog(TableEnvironmentImpl.java:337) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$null$5(ExecutionContext.java:627) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at java.util.HashMap.forEach(HashMap.java:1289) ~[?:1.8.0_251] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$initializeCatalogs$6(ExecutionContext.java:625) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:264) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:624) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:523) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:183) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:136) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:859) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.openSession(LocalExecutor.java:227) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:108) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > 2020-10-27 09:48:15,247 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Connected > to metastore. > 2020-10-27 09:48:15,247 INFO > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient [] - > RetryingMetaStoreClient proxy=class > org.apache.hadoop.hive.metastore.HiveMetaStoreClient ugi=hive (auth:SIMPLE) > retries=1 delay=1 lifetime=0 > 2020-10-27 09:48:15,364 WARN > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient [] - > MetaStoreClient lost connection. Attempting to reconnect (1 of 1) after 1s. > getDatabase > org.apache.thrift.transport.TTransportException: null > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:1135) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:1122) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1511) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1506) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[?:1.8.0_251] > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > ~[?:1.8.0_251] > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > ~[?:1.8.0_251] > at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_251] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:208) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at com.sun.proxy.$Proxy28.getDatabase(Unknown Source) ~[?:?] > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.getDatabase(HiveMetastoreClientWrapper.java:107) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:330) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.HiveCatalog.open(HiveCatalog.java:227) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.CatalogManager.registerCatalog(CatalogManager.java:191) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.registerCatalog(TableEnvironmentImpl.java:337) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$null$5(ExecutionContext.java:627) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at java.util.HashMap.forEach(HashMap.java:1289) ~[?:1.8.0_251] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$initializeCatalogs$6(ExecutionContext.java:625) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:264) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:624) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:523) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:183) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:136) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:859) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.openSession(LocalExecutor.java:227) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:108) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > 2020-10-27 09:48:16,365 INFO > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient [] - > RetryingMetaStoreClient trying reconnect as hive (auth:SIMPLE) > 2020-10-27 09:48:16,375 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Closed a > connection to metastore, current connections: 0 > 2020-10-27 09:48:16,375 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Trying to > connect to metastore with URI thrift://x.x.x.x:10000 > 2020-10-27 09:48:16,376 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Opened a > connection to metastore, current connections: 1 > 2020-10-27 09:48:16,436 WARN > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - set_ugi() > not successful, Likely cause: new client talking to old server. Continuing > without it. > org.apache.thrift.transport.TTransportException: null > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:380) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:230) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_set_ugi(ThriftHiveMetastore.java:4787) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.set_ugi(ThriftHiveMetastore.java:4773) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:534) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.reconnect(HiveMetaStoreClient.java:379) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient$1.run(RetryingMetaStoreClient.java:187) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at java.security.AccessController.doPrivileged(Native Method) > ~[?:1.8.0_251] > at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_251] > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836) > ~[flink-shaded-hadoop-2-uber-2.8.3-10.0.jar:2.8.3-10.0] > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:183) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at com.sun.proxy.$Proxy28.getDatabase(Unknown Source) ~[?:?] > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.getDatabase(HiveMetastoreClientWrapper.java:107) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:330) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.hive.HiveCatalog.open(HiveCatalog.java:227) > ~[flink-sql-connector-hive-3.1.2_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.catalog.CatalogManager.registerCatalog(CatalogManager.java:191) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.registerCatalog(TableEnvironmentImpl.java:337) > ~[flink-table_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$null$5(ExecutionContext.java:627) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at java.util.HashMap.forEach(HashMap.java:1289) ~[?:1.8.0_251] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$initializeCatalogs$6(ExecutionContext.java:625) > ~[flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:264) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:624) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:523) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:183) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:136) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:859) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.openSession(LocalExecutor.java:227) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:108) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201) > [flink-sql-client_2.12-1.11.1.jar:1.11.1] > 2020-10-27 09:48:16,438 INFO > org.apache.hadoop.hive.metastore.HiveMetaStoreClient [] - Connected > to metastore. > > > > > Exception in thread "main" > org.apache.flink.table.client.SqlClientException: Unexpected exception. > This is a bug. Please consider filing an issue. > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:213) > Caused by: org.apache.flink.table.client.gateway.SqlExecutionException: > Could not create execution context. > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:870) > at > org.apache.flink.table.client.gateway.local.LocalExecutor.openSession(LocalExecutor.java:227) > at org.apache.flink.table.client.SqlClient.start(SqlClient.java:108) > at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201) > Caused by: org.apache.flink.table.catalog.exceptions.CatalogException: > Failed to determine whether database default exists or not > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:335) > at > org.apache.flink.table.catalog.hive.HiveCatalog.open(HiveCatalog.java:227) > at > org.apache.flink.table.catalog.CatalogManager.registerCatalog(CatalogManager.java:191) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.registerCatalog(TableEnvironmentImpl.java:337) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$null$5(ExecutionContext.java:627) > at java.util.HashMap.forEach(HashMap.java:1289) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.lambda$initializeCatalogs$6(ExecutionContext.java:625) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:264) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:624) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:523) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:183) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.<init>(ExecutionContext.java:136) > at > org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:859) > ... 3 more > Caused by: org.apache.thrift.transport.TTransportException > at > org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132) > at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) > at > org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429) > at > org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318) > at > org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219) > at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_database(ThriftHiveMetastore.java:1135) > at > org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_database(ThriftHiveMetastore.java:1122) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1511) > at > org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getDatabase(HiveMetaStoreClient.java:1506) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:208) > at com.sun.proxy.$Proxy28.getDatabase(Unknown Source) > at > org.apache.flink.table.catalog.hive.client.HiveMetastoreClientWrapper.getDatabase(HiveMetastoreClientWrapper.java:107) > at > org.apache.flink.table.catalog.hive.HiveCatalog.databaseExists(HiveCatalog.java:330) > ... 15 more > > > 谢谢! > > > > -- Best regards! Rui Li
