[
https://issues.apache.org/jira/browse/FLINK-24942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
JasonLee closed FLINK-24942.
----------------------------
Resolution: Fixed
> Could not find any factory for identifier 'hive' that implements
> 'org.apache.flink.table.factories.DynamicTableFactory' in the classpath
> ----------------------------------------------------------------------------------------------------------------------------------------
>
> Key: FLINK-24942
> URL: https://issues.apache.org/jira/browse/FLINK-24942
> Project: Flink
> Issue Type: Bug
> Components: Table SQL / Client
> Affects Versions: 1.14.0
> Environment: Flink-1.14.0
> Reporter: JasonLee
> Priority: Major
> Fix For: 1.15.0
>
>
> [ERROR] Could not execute SQL statement. Reason:
> org.apache.flink.table.api.ValidationException: Could not find any factory
> for identifier 'hive' that implements
> 'org.apache.flink.table.factories.DynamicTableFactory' in the classpath.
> Available factory identifiers are:
> blackhole
> datagen
> filesystem
> kafka
> print
> upsert-kafka
>
> The above exception is thrown when I execute the following SQL, even though I
> have added flink-sql-connector-hive-2.3.6_2.11-1.14.0.jar in flink/lib
> {code:java}
> // code placeholder
> insert into fs_table
> select xxx,
> xxx,
> xxx,
> xxx,
> xxx,
> DATE_FORMAT(ts_ltz, 'yyyy-MM-dd'), DATE_FORMAT(ts_ltz, 'HH')
> from kafka_table; {code}
>
--
This message was sent by Atlassian Jira
(v8.20.1#820001)