[ 
https://issues.apache.org/jira/browse/FLINK-31659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17706726#comment-17706726
 ] 

luoyuxia commented on FLINK-31659:
----------------------------------

Not extactly,  it's not whenever as mentioned  "this bug only happens with 
filesystem tables. Other tables do not have this issue" by the reporter.

Back to this issue, the reason is we can't find the {{DialecFactory}} in 
FlinkUserClassloader and then we throw the excpetion. Also, I found 

we will catch the {{NoClassDefFoundError}} in 
{{{}FactoryUtil#discoverFactories,{}}}the context that why we catch 
{{NoClassDefFoundError}} can be found in 
[https://github.com/apache/flink/pull/17897.|https://github.com/apache/flink/pull/17897]

"when users bring in a format module, implementing 
{{{}BulkReaderFormatFactory{}}}/{{{}BulkWriterFormatFactory{}}}, but don't have 
flink-connector-files in the classpath because they don't use it. E.g. you're 
using {{flink-connector-kafka}} and you want to use {{{}flink-avro{}}}, which 
brings in an implementation of {{{}BulkWriterFormatFactory{}}}. "

The reason of the exception reported in this Jira is similar, we bring  Hive 
connector as well as  {{HiveDialectFactory}} . So, in method 
{{formatFactoryExists#formatFactoryExists}} it will try to load 
{{org.apache.flink.table.planner.delegation.hive.HiveDialectFactory}} by java 
SPI, but can't find the class defination. So the exception throws.

To fix this issue, I would like to catch the {{NoClassDefFoundError}} just like 
what we did in {{{}FactoryUtil#discoverFactories{}}}.

But it'll be just a fix to 1.17 & 1.16 for I'm intended to decouple Hive with 
planner in 1.18, and after that the problem won't happen.

> java.lang.ClassNotFoundException: 
> org.apache.flink.table.planner.delegation.DialectFactory when bundled Hive 
> connector jar is in classpath
> ------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: FLINK-31659
>                 URL: https://issues.apache.org/jira/browse/FLINK-31659
>             Project: Flink
>          Issue Type: Bug
>          Components: Connectors / Hive
>    Affects Versions: 1.17.0, 1.16.1, 1.15.4
>            Reporter: Caizhi Weng
>            Priority: Major
>
> Steps to reproduce this bug:
> 1. Download a fresh 1.16.1 Flink distribution.
> 2. Add both {{flink-sql-connector-hive-2.3.9_2.12-1.16.0.jar}} and 
> {{flink-shaded-hadoop-2-uber-2.8.3-10.0.jar}} into {{lib}} directory.
> 3. Start a standalone cluster with {{bin/start-cluster.sh}}.
> 4. Start SQL client and run the following SQL.
> {code:sql}
> create table T (
>   a int,
>   b string
> ) with (
>   'connector' = 'filesystem',
>   'format' = 'csv',
>   'path' = '/tmp/gao.csv'
> );
> create table S (
>   a int,
>   b string
> ) with (
>   'connector' = 'print'
> );
> insert into S select * from T;
> {code}
> The following exception will occur.
> {code}
> org.apache.flink.table.client.gateway.SqlExecutionException: Failed to parse 
> statement: insert into S select * from T;
>       at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:174)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.cli.SqlCommandParserImpl.parseCommand(SqlCommandParserImpl.java:45)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.cli.SqlMultiLineParser.parse(SqlMultiLineParser.java:71)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964) 
> ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778) 
> ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679) 
> ~[flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:295)
>  [flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:280)
>  [flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:228)
>  [flink-sql-client-1.16.1.jar:1.16.1]
>       at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151) 
> [flink-sql-client-1.16.1.jar:1.16.1]
>       at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95) 
> [flink-sql-client-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187) 
> [flink-sql-client-1.16.1.jar:1.16.1]
>       at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161) 
> [flink-sql-client-1.16.1.jar:1.16.1]
> Caused by: org.apache.flink.table.api.ValidationException: Unable to create a 
> source for reading table 'default_catalog.default_database.T'.
> Table options are:
> 'connector'='filesystem'
> 'format'='json'
> 'path'='/tmp/gao.json'
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:166)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106)
>  ~[?:?]
>       at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       ... 12 more
> Caused by: java.lang.NoClassDefFoundError: 
> org/apache/flink/table/planner/delegation/DialectFactory
>       at java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_361]
>       at java.lang.ClassLoader.defineClass(ClassLoader.java:756) 
> ~[?:1.8.0_361]
>       at 
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.access$100(URLClassLoader.java:74) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_361]
>       at java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_361]
>       at java.security.AccessController.doPrivileged(Native Method) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.findClass(URLClassLoader.java:362) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361]
>       at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:405) ~[?:1.8.0_361]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at java.lang.Class.forName0(Native Method) ~[?:1.8.0_361]
>       at java.lang.Class.forName(Class.java:348) ~[?:1.8.0_361]
>       at 
> java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) 
> ~[?:1.8.0_361]
>       at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) 
> ~[?:1.8.0_361]
>       at java.util.ServiceLoader$1.next(ServiceLoader.java:480) ~[?:1.8.0_361]
>       at java.util.Iterator.forEachRemaining(Iterator.java:116) ~[?:1.8.0_361]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.formatFactoryExists(FileSystemTableFactory.java:206)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.discoverDecodingFormat(FileSystemTableFactory.java:172)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.createDynamicTableSource(FileSystemTableFactory.java:77)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:163)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106)
>  ~[?:?]
>       at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       ... 12 more
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.flink.table.planner.delegation.DialectFactory
>       at java.net.URLClassLoader.findClass(URLClassLoader.java:387) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361]
>       at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361]
>       at java.lang.ClassLoader.defineClass1(Native Method) ~[?:1.8.0_361]
>       at java.lang.ClassLoader.defineClass(ClassLoader.java:756) 
> ~[?:1.8.0_361]
>       at 
> java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.access$100(URLClassLoader.java:74) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader$1.run(URLClassLoader.java:369) ~[?:1.8.0_361]
>       at java.net.URLClassLoader$1.run(URLClassLoader.java:363) ~[?:1.8.0_361]
>       at java.security.AccessController.doPrivileged(Native Method) 
> ~[?:1.8.0_361]
>       at java.net.URLClassLoader.findClass(URLClassLoader.java:362) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_361]
>       at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355) 
> ~[?:1.8.0_361]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:405) ~[?:1.8.0_361]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:67)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:51)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_361]
>       at 
> org.apache.flink.util.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:192)
>  ~[flink-dist-1.16.1.jar:1.16.1]
>       at java.lang.Class.forName0(Native Method) ~[?:1.8.0_361]
>       at java.lang.Class.forName(Class.java:348) ~[?:1.8.0_361]
>       at 
> java.util.ServiceLoader$LazyIterator.nextService(ServiceLoader.java:370) 
> ~[?:1.8.0_361]
>       at java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:404) 
> ~[?:1.8.0_361]
>       at java.util.ServiceLoader$1.next(ServiceLoader.java:480) ~[?:1.8.0_361]
>       at java.util.Iterator.forEachRemaining(Iterator.java:116) ~[?:1.8.0_361]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.formatFactoryExists(FileSystemTableFactory.java:206)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.discoverDecodingFormat(FileSystemTableFactory.java:172)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.connector.file.table.FileSystemTableFactory.createDynamicTableSource(FileSystemTableFactory.java:77)
>  ~[flink-connector-files-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:163)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191)
>  ~[flink-table-api-java-uber-1.16.1.jar:1.16.1]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462)
>  ~[?:?]
>       at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNodeOrFail(SqlToOperationConverter.java:384)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:828)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:351)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262)
>  ~[?:?]
>       at 
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106)
>  ~[?:?]
>       at 
> org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172)
>  ~[flink-sql-client-1.16.1.jar:1.16.1]
>       ... 12 more
> {code}
> The 
> [document|https://nightlies.apache.org/flink/flink-docs-release-1.16/docs/connectors/table/hive/overview/#moving-the-planner-jar]
>  states that if we're not using Hive dialect, we don't have to move the 
> planner jar. However we clearly did not use Hive dialect in the above SQL, 
> but the exception still happens.
> This bug only happens with filesystem tables. Other tables do not have this 
> issue.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to