[jira] [Commented] (FLINK-28903) flink-table-store-hive-catalog could not shade hive-shims-0.23
[ https://issues.apache.org/jira/browse/FLINK-28903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17578291#comment-17578291 ] Nicholas Jiang commented on FLINK-28903: [~lzljs3620320] , when the value of hive.metastore.use.SSL or hive.metastore.sasl.enabled is true, there is above exception for Hive 2.x including Hive 2.3. > flink-table-store-hive-catalog could not shade hive-shims-0.23 > -- > > Key: FLINK-28903 > URL: https://issues.apache.org/jira/browse/FLINK-28903 > Project: Flink > Issue Type: Bug > Components: Table Store >Affects Versions: table-store-0.3.0 >Reporter: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: table-store-0.3.0 > > > flink-table-store-hive-catalog could not shade hive-shims-0.23 because > artifactSet doesn't include hive-shims-0.23 and the minimizeJar is set to > true. The exception is as follows: > {code:java} > Caused by: java.lang.RuntimeException: Unable to instantiate > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.HiveMetaStoreClient > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1708) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:97) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalog.createClient(HiveCatalog.java:380) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalog.(HiveCatalog.java:80) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalogFactory.create(HiveCatalogFactory.java:51) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.file.catalog.CatalogFactory.createCatalog(CatalogFactory.java:93) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:62) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:57) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:31) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.factories.FactoryUtil.createCatalog(FactoryUtil.java:428) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.createCatalog(TableEnvironmentImpl.java:1356) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$executeOperation$3(LocalExecutor.java:209) > ~[flink-sql-client-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.context.ExecutionContext.wrapClassLoader(ExecutionContext.java:88) > ~[flink-sql-client-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.executeOperation(LocalExecutor.java:209) > ~[flink-sql-client-1.15.1.jar:1.15.1] > ... 10 more > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > ~[?:1.8.0_181] > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > ~[?:1.8.0_181] > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > ~[?:1.8.0_181] > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > ~[?:1.8.0_181] > at >
[jira] [Commented] (FLINK-28903) flink-table-store-hive-catalog could not shade hive-shims-0.23
[ https://issues.apache.org/jira/browse/FLINK-28903?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17578248#comment-17578248 ] Jingsong Lee commented on FLINK-28903: -- We can change this to `support hive 2.1&2.2` > flink-table-store-hive-catalog could not shade hive-shims-0.23 > -- > > Key: FLINK-28903 > URL: https://issues.apache.org/jira/browse/FLINK-28903 > Project: Flink > Issue Type: Bug > Components: Table Store >Affects Versions: table-store-0.3.0 >Reporter: Nicholas Jiang >Priority: Major > Labels: pull-request-available > Fix For: table-store-0.3.0 > > > flink-table-store-hive-catalog could not shade hive-shims-0.23 because > artifactSet doesn't include hive-shims-0.23 and the minimizeJar is set to > true. The exception is as follows: > {code:java} > Caused by: java.lang.RuntimeException: Unable to instantiate > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.HiveMetaStoreClient > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1708) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.(RetryingMetaStoreClient.java:83) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:133) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:104) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.getProxy(RetryingMetaStoreClient.java:97) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalog.createClient(HiveCatalog.java:380) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalog.(HiveCatalog.java:80) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.hive.HiveCatalogFactory.create(HiveCatalogFactory.java:51) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.file.catalog.CatalogFactory.createCatalog(CatalogFactory.java:93) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:62) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:57) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.store.connector.FlinkCatalogFactory.createCatalog(FlinkCatalogFactory.java:31) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at > org.apache.flink.table.factories.FactoryUtil.createCatalog(FactoryUtil.java:428) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.createCatalog(TableEnvironmentImpl.java:1356) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:) > ~[flink-table-api-java-uber-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$executeOperation$3(LocalExecutor.java:209) > ~[flink-sql-client-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.context.ExecutionContext.wrapClassLoader(ExecutionContext.java:88) > ~[flink-sql-client-1.15.1.jar:1.15.1] > at > org.apache.flink.table.client.gateway.local.LocalExecutor.executeOperation(LocalExecutor.java:209) > ~[flink-sql-client-1.15.1.jar:1.15.1] > ... 10 more > Caused by: java.lang.reflect.InvocationTargetException > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > ~[?:1.8.0_181] > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > ~[?:1.8.0_181] > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > ~[?:1.8.0_181] > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > ~[?:1.8.0_181] > at > org.apache.flink.table.store.shaded.org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1706) > ~[flink-table-store-dist-0.3-SNAPSHOT.jar:0.3-SNAPSHOT] > at >