我看了下文档,几个包都在,还有一个单独下载依赖的包flink-sql-connector-hive-3.1.3,不知道是使用这个还是下面的?
// Flink's Hive connector  flink-connector-hive_2.12-1.19.1.jar  // Hive 
dependencies  hive-exec-3.1.0.jar  libfb303-0.9.3.jar // libfb303 is not packed 
into hive-exec in some versions, need to add it separately  // add 
antlr-runtime if you need to use hive dialect  antlr-runtime-3.5.2.jar
lib下面的包
antlr-runtime-3.5.2.jar flink-table-api-java-1.19.0.jar 
flink-cdc-dist-3.0.0.jar flink-table-api-java-uber-1.19.0.jar 
flink-cdc-pipeline-connector-doris-3.1.0.jar flink-table-common-1.19.0.jar 
flink-cdc-pipeline-connector-mysql-3.1.0.jar 
flink-table-planner-loader-1.19.0.jar flink-cep-1.19.0.jar 
flink-table-runtime-1.19.0.jar flink-connector-files-1.19.0.jar 
hive-exec-3.1.2.jar flink-connector-hive_2.12-1.19.0.jar libfb303-0.9.3.jar 
flink-connector-jdbc-3.1.2-1.18.jar log4j-1.2-api-2.17.1.jar 
flink-connector-kafka-3.1.0-1.18.jar log4j-api-2.17.1.jar flink-csv-1.19.0.jar 
log4j-core-2.17.1.jar flink-dist-1.19.0.jar log4j-slf4j-impl-2.17.1.jar 
flink-json-1.19.0.jar mysql-connector-java-8.0.28.jar 
flink-scala_2.12-1.19.0.jar paimon-flink-1.19-0.9-20240628.002224-23.jar
------------------------------------------------------------------
发件人:Xuyang <xyzhong...@163.com>
发送时间:2024年7月16日(星期二) 11:43
收件人:"user-zh"<user-zh@flink.apache.org>
主 题:Re:使用hive的catalog问题
Hi, 可以check一下是否将hive sql connector的依赖[1]放入lib目录下或者add jar了吗?
[1] 
https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/overview/
 
<https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/overview/
 >
--
 Best!
 Xuyang
At 2024-07-15 17:09:45, "冯奇" <ha.fen...@aisino.com> wrote:
>Flink SQL> USE CATALOG myhive;
>Flink SQL> CREATE TABLE mykafka (name String, age Int) WITH (
> 'connector.type' = 'kafka',
> 'connector.version' = 'universal',
> 'connector.topic' = 'hive_sink',
> 'connector.properties.bootstrap.servers' = '10.0.15.242:9092',
> 'format.type' = 'csv',
> 'update-mode' = 'append'
>);
>提示下面错误:
>[ERROR] Could not execute SQL statement. Reason:
>org.apache.flink.table.factories.NoMatchingTableFactoryException: Could not 
>find a suitable table factory for 
>'org.apache.flink.table.factories.TableSourceFactory' in
>the classpath.
>Reason: Required context properties mismatch.
>The matching candidates:
>org.apache.flink.table.sources.CsvAppendTableSourceFactory
>Mismatched properties:
>'connector.type' expects 'filesystem', but is 'kafka'
>The following properties are requested:
>connector.properties.bootstrap.servers=10.0.15.242:9092
>connector.topic=hive_sink
>connector.type=kafka
>connector.version=universal
>format.type=csv
>schema.0.data-type=VARCHAR(2147483647)
>schema.0.name=name
>schema.1.data-type=INT
>schema.1.name=age
>update-mode=append
>The following factories have been considered:
>org.apache.flink.table.sources.CsvBatchTableSourceFactory
>org.apache.flink.table.sources.CsvAppendTableSourceFactory

回复