这样试试,把”\”改成”/“: file:///D:/Pyproject/flink-connector-jdbc_2.11-1.13.1.jar
> 2021年6月1日 下午5:40,琴师 <1129656...@qq.com> 写道: > > 再请叫一个问题,我在pycharm使用时候引用windows的地址不能引用,比如file:///D:\Pyproject\flink-connector-jdbc_2.11-1.13.1.jar; > 这样不能引入,大佬有用用过吗? > > > ------------------ 原始邮件 ------------------ > 发件人: > "琴师" > > <1129656...@qq.com>; > 发送时间: 2021年6月1日(星期二) 下午5:30 > 收件人: "user-zh"<user-zh@flink.apache.org>; > > 主题: 回复: Pyflink jdbc相关 > > > > > > 感谢,我换成2.11确实可以了!!!! > > > ------------------ 原始邮件 ------------------ > 发件人: > "user-zh" > > <dian0511...@gmail.com>; > 发送时间: 2021年6月1日(星期二) 下午5:04 > 收件人: "user-zh"<user-zh@flink.apache.org>;"琴师"<1129656...@qq.com>; > > 主题: Re: Pyflink jdbc相关 > > > > Hi, > > 本地执行: > 1)用flink-connector-jdbc_2.11-1.13.1.jar试试?因为PyFlink里默认待的JAR包是scala 2.11的 > > > flink run: > 1) 你注册的sink表的名字为“print”,不是”table_sink”,但是在SQL语句里用的table_sink。 > > > > 2021年6月1日 下午4:33,琴师 <1129656...@qq.com> 写道: > > > > Hi, > > &nbsp; > &nbsp;我按着微信分享https://mp.weixin.qq.com/s/GyFTjQl6ch8jc733mpCP7Q > 试着使用pyflink,遇到了问题,我引用了jdbc的jar包,但是仍然提示我jdbc错误。我的flink版本是1.13.1 > > 我的原代码如下: > > > > > > from pyflink.datastream import StreamExecutionEnvironment > > from pyflink.table import StreamTableEnvironment, EnvironmentSettings > > env = StreamExecutionEnvironment.get_execution_environment() > > table_env = StreamTableEnvironment.create( > env,environment_settings=EnvironmentSettings.new_instance().use_blink_planner().build()) > > > table_env.get_config().get_configuration().set_string("pipeline.classpaths", > "file:///home/flink/lib/flink-connector-jdbc_2.12-1.13.1.jar;file:///home/flink/lib/mysql-connector-java-8.0.25.jar") > > > > > > # 2. create source Table > > table_env.execute_sql(""" > > > > > > CREATE TABLE table_source ( > > &nbsp; e string > > ) WITH ( > > &nbsp;'connector' = 'jdbc', > > &nbsp; 'url' = 'jdbc:mysql://********:3306/test', > > &nbsp; 'driver' = 'com.mysql.cj.jdbc.Driver', > > &nbsp; 'table-name' = 'enum_test', > > &nbsp; 'username' = 'pms_etl', > > &nbsp; 'password' = 'pms_etl_q' > > ) > > > > > > """) > > > > > > # 3. create sink Table > > table_env.execute_sql(""" > > &nbsp; &nbsp; CREATE TABLE print ( > > &nbsp; &nbsp; &nbsp; &nbsp; e string > > &nbsp; &nbsp; ) WITH ( > > &nbsp; &nbsp; &nbsp; &nbsp; 'connector' = 'print' > > &nbsp; &nbsp; ) > > """) > > &nbsp; &nbsp; > > > > > > table_env.execute_sql("INSERT INTO table_sink SELECT * FROM > table_source").wait() > > > > > > > > 我直接用python执行时候错误返回如下 > > > > > > Traceback (most recent call last): > > &nbsp; File "demo.py", line 41, in <module&gt; > > &nbsp; &nbsp; table_env.execute_sql("INSERT INTO table_sink > SELECT * FROM table_source").wait() > > &nbsp; File > "/usr/local/lib/python3.7/site-packages/pyflink/table/table_environment.py", > line 804, in execute_sql > > &nbsp; &nbsp; return TableResult(self._j_tenv.executeSql(stmt)) > > &nbsp; File > "/usr/local/lib/python3.7/site-packages/py4j/java_gateway.py", line 1286, in > __call__ > > &nbsp; &nbsp; answer, self.gateway_client, self.target_id, > self.name) > > &nbsp; File > "/usr/local/lib/python3.7/site-packages/pyflink/util/exceptions.py", line > 146, in deco > > &nbsp; &nbsp; return f(*a, **kw) > > &nbsp; File > "/usr/local/lib/python3.7/site-packages/py4j/protocol.py", line 328, in > get_return_value > > &nbsp; &nbsp; format(target_id, ".", name), value) > > py4j.protocol.Py4JJavaError: An error occurred while calling > o4.executeSql. > > : org.apache.flink.table.api.ValidationException: Unable to create a > source for reading table 'default_catalog.default_database.table_source'. > > > > > > Table options are: > > > > > > 'connector'='jdbc' > > 'driver'='com.mysql.cj.jdbc.Driver' > > 'password'='pms_etl_q' > > 'table-name'='enum_test' > > 'url'='jdbc:mysql://*******:3306/test' > > 'username'='pms_etl' > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.createTableSource(FactoryUtil.java:137) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:116) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:82) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3585) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2507) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2144) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2093) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2050) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:663) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:644) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3438) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:170) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:162) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:967) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:936) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:275) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlInsert(SqlToOperationConverter.java:595) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:268) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:101) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeSql(TableEnvironmentImpl.java:724) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > &nbsp; &nbsp; &nbsp; &nbsp; at > java.lang.reflect.Method.invoke(Method.java:498) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) > > &nbsp; &nbsp; &nbsp; &nbsp; at > java.lang.Thread.run(Thread.java:748) > > Caused by: org.apache.flink.table.api.ValidationException: Cannot > discover a connector using option: 'connector'='jdbc' > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.enrichNoMatchingConnectorError(FactoryUtil.java:467) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.getDynamicTableFactory(FactoryUtil.java:441) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.createTableSource(FactoryUtil.java:133) > > &nbsp; &nbsp; &nbsp; &nbsp; ... 31 more > > Caused by: org.apache.flink.table.api.ValidationException: Could not > find any factory for identifier 'jdbc' that implements > 'org.apache.flink.table.factories.DynamicTableFactory' in the classpath. > > > > > > Available factory identifiers are: > > > > > > blackhole > > datagen > > filesystem > > print > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.discoverFactory(FactoryUtil.java:319) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.factories.FactoryUtil.enrichNoMatchingConnectorError(FactoryUtil.java:463) > > &nbsp; &nbsp; &nbsp; &nbsp; ... 33 more > > > > > > > > 我用flink run -py demo.py 返回错误如下: > > > > > > &nbsp; File "./demo.py", line 41, in <module&gt; > > &nbsp; &nbsp; table_env.execute_sql("INSERT INTO table_sink > SELECT * FROM table_source").wait() > > &nbsp; File > "/home/flink/opt/python/pyflink.zip/pyflink/table/table_environment.py", line > 804, in execute_sql > > &nbsp; File > "/home/flink/opt/python/py4j-0.10.8.1-src.zip/py4j/java_gateway.py", line > 1286, in __call__ > > &nbsp; File > "/home/flink/opt/python/pyflink.zip/pyflink/util/exceptions.py", line 158, in > deco > > pyflink.util.exceptions.TableException: > org.apache.flink.table.api.TableException: Sink > `default_catalog`.`default_database`.`table_sink` does not exists > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.delegation.PlannerBase.translateToRel(PlannerBase.scala:233) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.delegation.PlannerBase.$anonfun$translate$1(PlannerBase.scala:162) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:233) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.Iterator.foreach(Iterator.scala:937) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.Iterator.foreach$(Iterator.scala:937) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.AbstractIterator.foreach(Iterator.scala:1425) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.IterableLike.foreach(IterableLike.scala:70) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.IterableLike.foreach$(IterableLike.scala:69) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.AbstractIterable.foreach(Iterable.scala:54) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.TraversableLike.map(TraversableLike.scala:233) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.TraversableLike.map$(TraversableLike.scala:226) > > &nbsp; &nbsp; &nbsp; &nbsp; at > scala.collection.AbstractTraversable.map(Traversable.scala:104) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:162) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:1518) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:740) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:856) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.table.api.internal.TableEnvironmentImpl.executeSql(TableEnvironmentImpl.java:730) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > &nbsp; &nbsp; &nbsp; &nbsp; at > java.lang.reflect.Method.invoke(Method.java:498) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.Gateway.invoke(Gateway.java:282) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.commands.CallCommand.execute(CallCommand.java:79) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.api.python.shaded.py4j.GatewayConnection.run(GatewayConnection.java:238) > > &nbsp; &nbsp; &nbsp; &nbsp; at > java.lang.Thread.run(Thread.java:748) > > > > > > org.apache.flink.client.program.ProgramAbortException: > java.lang.RuntimeException: Python process exits with code: 1 > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:134) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > &nbsp; &nbsp; &nbsp; &nbsp; at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > &nbsp; &nbsp; &nbsp; &nbsp; at > java.lang.reflect.Method.invoke(Method.java:498) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:355) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:222) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:812) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:246) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1054) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28) > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132) > > Caused by: java.lang.RuntimeException: Python process exits with code: 1 > > &nbsp; &nbsp; &nbsp; &nbsp; at > org.apache.flink.client.python.PythonDriver.main(PythonDriver.java:124) > > &nbsp; &nbsp; &nbsp; &nbsp; ... 13 more > > > > > > > > 请问我该如何解决?