上面的示例好像使用的旧版本的 kafka connector 参数。 参考文档使用新版本的参数: https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/hive_catalog/#step-4-create-a-kafka-table-with-flink-sql-ddl 需要把 kafka 的 connector [1] 也放入到 lib 目录下。
[1] https://nightlies.apache.org/flink/flink-docs-release-1.19/docs/connectors/table/kafka/ Best, Feng On Tue, Jul 16, 2024 at 2:11 PM Xuyang <xyzhong...@163.com> wrote: > lib目录下,需要放置一下flink-sql-connector-hive-3.1.3,这个包是给sql作业用的 > > > > > -- > > Best! > Xuyang > > > > > > 在 2024-07-16 13:40:23,"冯奇" <ha.fen...@aisino.com> 写道: > >我看了下文档,几个包都在,还有一个单独下载依赖的包flink-sql-connector-hive-3.1.3,不知道是使用这个还是下面的? > >// Flink's Hive connector flink-connector-hive_2.12-1.19.1.jar // Hive > dependencies hive-exec-3.1.0.jar libfb303-0.9.3.jar // libfb303 is not > packed into hive-exec in some versions, need to add it separately // add > antlr-runtime if you need to use hive dialect antlr-runtime-3.5.2.jar > >lib下面的包 > >antlr-runtime-3.5.2.jar flink-table-api-java-1.19.0.jar > flink-cdc-dist-3.0.0.jar flink-table-api-java-uber-1.19.0.jar > flink-cdc-pipeline-connector-doris-3.1.0.jar flink-table-common-1.19.0.jar > flink-cdc-pipeline-connector-mysql-3.1.0.jar > flink-table-planner-loader-1.19.0.jar flink-cep-1.19.0.jar > flink-table-runtime-1.19.0.jar flink-connector-files-1.19.0.jar > hive-exec-3.1.2.jar flink-connector-hive_2.12-1.19.0.jar libfb303-0.9.3.jar > flink-connector-jdbc-3.1.2-1.18.jar log4j-1.2-api-2.17.1.jar > flink-connector-kafka-3.1.0-1.18.jar log4j-api-2.17.1.jar > flink-csv-1.19.0.jar log4j-core-2.17.1.jar flink-dist-1.19.0.jar > log4j-slf4j-impl-2.17.1.jar flink-json-1.19.0.jar > mysql-connector-java-8.0.28.jar flink-scala_2.12-1.19.0.jar > paimon-flink-1.19-0.9-20240628.002224-23.jar > >------------------------------------------------------------------ > >发件人:Xuyang <xyzhong...@163.com> > >发送时间:2024年7月16日(星期二) 11:43 > >收件人:"user-zh"<user-zh@flink.apache.org> > >主 题:Re:使用hive的catalog问题 > >Hi, 可以check一下是否将hive sql connector的依赖[1]放入lib目录下或者add jar了吗? > >[1] > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/overview/ > < > https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/overview/ > > > >-- > > Best! > > Xuyang > >At 2024-07-15 17:09:45, "冯奇" <ha.fen...@aisino.com> wrote: > >>Flink SQL> USE CATALOG myhive; > >>Flink SQL> CREATE TABLE mykafka (name String, age Int) WITH ( > >> 'connector.type' = 'kafka', > >> 'connector.version' = 'universal', > >> 'connector.topic' = 'hive_sink', > >> 'connector.properties.bootstrap.servers' = '10.0.15.242:9092', > >> 'format.type' = 'csv', > >> 'update-mode' = 'append' > >>); > >>提示下面错误: > >>[ERROR] Could not execute SQL statement. Reason: > >>org.apache.flink.table.factories.NoMatchingTableFactoryException: Could > not find a suitable table factory for > 'org.apache.flink.table.factories.TableSourceFactory' in > >>the classpath. > >>Reason: Required context properties mismatch. > >>The matching candidates: > >>org.apache.flink.table.sources.CsvAppendTableSourceFactory > >>Mismatched properties: > >>'connector.type' expects 'filesystem', but is 'kafka' > >>The following properties are requested: > >>connector.properties.bootstrap.servers=10.0.15.242:9092 > >>connector.topic=hive_sink > >>connector.type=kafka > >>connector.version=universal > >>format.type=csv > >>schema.0.data-type=VARCHAR(2147483647) > >>schema.0.name=name > >>schema.1.data-type=INT > >>schema.1.name=age > >>update-mode=append > >>The following factories have been considered: > >>org.apache.flink.table.sources.CsvBatchTableSourceFactory > >>org.apache.flink.table.sources.CsvAppendTableSourceFactory >