[ 
https://issues.apache.org/jira/browse/FLINK-30318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17645907#comment-17645907
 ] 

Samrat Deb commented on FLINK-30318:
------------------------------------

[~lsy] , m on it ! i will share the root cause once i find something. 

There is one more interesting thing i can see 

Add jar command from cli does not loads the custom jar , (i have added some 
stack traces, thread dumps and enabled debug logs to gain more insights ). 
{code:java}
Flink SQL> ADD JAR '/home/hadoop/flink-connector-mysql-cdc-2.3.0.jar';
Hello Here----
java.lang.RuntimeException: e
    at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.addJar(TableEnvironmentImpl.java:473)
    at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:1177)
    at 
org.apache.flink.table.client.gateway.local.LocalExecutor.executeOperation(LocalExecutor.java:206)
    at 
org.apache.flink.table.client.cli.CliClient.executeOperation(CliClient.java:639)
    at 
org.apache.flink.table.client.cli.CliClient.callOperation(CliClient.java:473)
    at 
org.apache.flink.table.client.cli.CliClient.executeOperation(CliClient.java:372)
    at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:329)
    at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:280)
    at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:228)
    at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151)
    at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95)
    at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187)
    at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161)
2022-12-12 04:25:39,787 INFO  org.apache.flink.table.resource.ResourceManager   
           [] - Added jar resource 
[file:/home/hadoop/flink-connector-mysql-cdc-2.3.0.jar] to class path.
2022-12-12 04:25:39,787 INFO  org.apache.flink.table.resource.ResourceManager   
           [] - Register resource 
[/home/hadoop/flink-connector-mysql-cdc-2.3.0.jar] successfully.
[INFO] Execute statement succeed.{code}
 

then show jar command 
{code:java}
Flink SQL> show jars;
+--------------------------------------------------+
|                                             jars |
+--------------------------------------------------+
| /home/hadoop/flink-connector-mysql-cdc-2.3.0.jar |
+--------------------------------------------------+
1 row in set{code}
 

creating a mysql cdc table 


{code:java}
Flink SQL> CREATE TABLE mysql_sample_table (
>       id INT,
>       name VARCHAR(50) NOT NULL,
>       PRIMARY KEY(id) NOT ENFORCED
>       ) WITH (
>       'connector' = 'mysql-cdc',
>       'hostname' = 'abcdcefgh.rds.amazonaws.com',
>       'port' = '3306',
>       'username' = 'user',
>       'password' = 'pass',
>       'database-name' = 'demo',
>       'table-name' = 'test_table'
>  );
[INFO] Execute statement succeed.{code}
then selecting from the table 



 
{code:java}
Flink SQL> select * from mysql_sample_table;
2022-12-12 04:26:30,302 WARN  org.apache.flink.table.client.cli.CliClient       
           [] - Could not execute SQL statement.
org.apache.flink.table.client.gateway.SqlExecutionException: Failed to parse 
statement: select * from mysql_sample_table;
    at 
org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:174)
 ~[flink-sql-client-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.client.cli.SqlCommandParserImpl.parseCommand(SqlCommandParserImpl.java:45)
 ~[flink-sql-client-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.client.cli.SqlMultiLineParser.parse(SqlMultiLineParser.java:71)
 ~[flink-sql-client-1.16.0.jar:1.16.0]
    at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2731) 
~[jline-3.9.0.jar:?]
    at org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:585) 
~[jline-3.9.0.jar:?]
    at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:295)
 [flink-sql-client-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:280)
 [flink-sql-client-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:228)
 [flink-sql-client-1.16.0.jar:1.16.0]
    at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:151) 
[flink-sql-client-1.16.0.jar:1.16.0]
    at org.apache.flink.table.client.SqlClient.start(SqlClient.java:95) 
[flink-sql-client-1.16.0.jar:1.16.0]
    at org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:187) 
[flink-sql-client-1.16.0.jar:1.16.0]
    at org.apache.flink.table.client.SqlClient.main(SqlClient.java:161) 
[flink-sql-client-1.16.0.jar:1.16.0]
Caused by: org.apache.flink.table.api.ValidationException: Unable to create a 
source for reading table 'default_catalog.default_database.mysql_sample_table'.
Table options are:
'connector'='mysql-cdc'
'database-name'='demo'
'hostname'='abcdefgh.rds.amazonaws.com'
'password'='******'
'port'='3306'
'table-name'='test_table'
'username'='user'
    at 
org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:166)
 ~[flink-table-api-java-uber-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.factories.FactoryUtil.createDynamicTableSource(FactoryUtil.java:191)
 ~[flink-table-api-java-uber-1.16.0.jar:1.16.0]
    at 
org.apache.flink.table.planner.plan.schema.CatalogSourceTable.createDynamicTableSource(CatalogSourceTable.java:175)
 ~[?:?]
    at 
org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.java:115)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3619) 
~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2559)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2175)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2095)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2038)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:669)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:657)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3462)
 ~[?:?]
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
 ~[?:?]
    at 
org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215)
 ~[?:?]
    at 
org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191)
 ~[?:?]
    at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498)
 ~[?:?]
~
~
"taskmanager.log" 309L, 78831B                                                  
            309,69        Bot
directory.info  launch_container.sh  prelaunch.err  prelaunch.out  
taskmanager.err  taskmanager.log  taskmanager.out
[root@ip-172-31-42-148 container_1670653965628_0004_01_000002]# cd ..
[root@ip-172-31-42-148 application_1670653965628_0004]# ls
container_1670653965628_0004_01_000001  container_1670653965628_0004_01_000002
[root@ip-172-31-42-148 application_1670653965628_0004]# cd ..
[root@ip-172-31-42-148 containers]# ls
application_1670653965628_0001  application_1670653965628_0002  
application_1670653965628_0003  application_1670653965628_0004
[root@ip-172-31-42-148 containers]# cd application_1670653965628_0004
[root@ip-172-31-42-148 application_1670653965628_0004]# ls
    at 
org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:570)
 ~[?:?]
    at 
org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:215)
 ~[?:?]
    at 
org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:191)
 ~[?:?]
    at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:1498)
 ~[?:?]
    at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:1253)
 ~[?:?]
    at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convertValidatedSqlNode(SqlToOperationConverter.java:374)
 ~[?:?]
    at 
org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:262)
 ~[?:?]
    at 
org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:106) 
~[?:?]
    at 
org.apache.flink.table.client.gateway.local.LocalExecutor.parseStatement(LocalExecutor.java:172)
 ~[flink-sql-client-1.16.0.jar:1.16.0]
    ... 11 more
[ERROR] Could not execute SQL statement. Reason:
java.lang.ClassNotFoundException: 
com.ververica.cdc.debezium.utils.ResolvedSchemaUtils
{code}
i tried similar activity on flink 1.15.2 , similar commands don't fail. 

One more thing is putting the custom jar in flink's lib path , it works for 
1.16.0 . 

 

> sql-client failed to load jar passed with -j args in flink 1.16.0 
> ------------------------------------------------------------------
>
>                 Key: FLINK-30318
>                 URL: https://issues.apache.org/jira/browse/FLINK-30318
>             Project: Flink
>          Issue Type: Bug
>          Components: Table SQL / Client
>    Affects Versions: 1.16.0
>            Reporter: Samrat Deb
>            Priority: Major
>
> In Flink 1.16.0 , /usr/lib/flink/bin/sql-client.sh embedded -j 
> hudi-flink-bundle_2.12-0.10.1.jar failed to load jar passed through argument 
>  
> {code:java}
>  /usr/lib/flink/bin/sql-client.sh embedded -j 
> hudi-flink-bundle_2.12-0.10.1.jar
> Setting HBASE_CONF_DIR=/etc/hbase/conf because no HBASE_CONF_DIR was set.
> Using the result of 'hadoop classpath' to augment the Hadoop classpath: 
> /etc/hadoop/conf:/usr/lib/hadoop/lib/:/usr/lib/hadoop/.//:/usr/lib/hadoop-hdfs/./:/usr/lib/hadoop-hdfs/lib/:/usr/lib/hadoop-hdfs/.//:/usr/lib/hadoop-mapreduce/.//:/usr/lib/hadoop-yarn/lib/:/usr/lib/hadoop-yarn/.//:/etc/tez/conf:/usr/lib/tez/hadoop-shim-0.10.2-amzn-0.jar:/usr/lib/tez/hadoop-shim-2.8-0.10.2-amzn-0.jar:/usr/lib/tez/lib:/usr/lib/tez/LICENSE:/usr/lib/tez/LICENSE-BSD-3clause:/usr/lib/tez/LICENSE-CDDLv1.1-GPLv2_withCPE:/usr/lib/tez/LICENSE-MIT:/usr/lib/tez/LICENSE-SIL_OpenFontLicense-v1.1:/usr/lib/tez/NOTICE:/usr/lib/tez/tez-api-0.10.2-amzn-0.jar:/usr/lib/tez/tez-aux-services-0.10.2-amzn-0.jar:/usr/lib/tez/tez-build-tools-0.10.2-amzn-0.jar:/usr/lib/tez/tez-common-0.10.2-amzn-0.jar:/usr/lib/tez/tez-dag-0.10.2-amzn-0.jar:/usr/lib/tez/tez-examples-0.10.2-amzn-0.jar:/usr/lib/tez/tez-history-parser-0.10.2-amzn-0.jar:/usr/lib/tez/tez-javadoc-tools-0.10.2-amzn-0.jar:/usr/lib/tez/tez-job-analyzer-0.10.2-amzn-0.jar:/usr/lib/tez/tez-mapreduce-0.10.2-amzn-0.jar:/usr/lib/tez/tez-protobuf-history-plugin-0.10.2-amzn-0.jar:/usr/lib/tez/tez-runtime-internals-0.10.2-amzn-0.jar:/usr/lib/tez/tez-runtime-library-0.10.2-amzn-0.jar:/usr/lib/tez/tez-tests-0.10.2-amzn-0.jar:/usr/lib/tez/tez-ui-0.10.2-amzn-0.war:/usr/lib/tez/tez-yarn-timeline-cache-plugin-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-with-acls-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-with-fs-0.10.2-amzn-0.jar:/usr/lib/tez/lib/async-http-client-2.12.3.jar:/usr/lib/tez/lib/commons-cli-1.2.jar:/usr/lib/tez/lib/commons-codec-1.11.jar:/usr/lib/tez/lib/commons-collections4-4.1.jar:/usr/lib/tez/lib/commons-io-2.8.0.jar:/usr/lib/tez/lib/commons-lang-2.6.jar:/usr/lib/tez/lib/guava-31.1-jre.jar:/usr/lib/tez/lib/hadoop-annotations.jar:/usr/lib/tez/lib/hadoop-auth.jar:/usr/lib/tez/lib/hadoop-hdfs-client-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-common-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-common.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-core-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-yarn-server-timeline-pluginstorage-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-yarn-server-web-proxy.jar:/usr/lib/tez/lib/hhadoop-mapreduce-client-core.jar:/usr/lib/tez/lib/javax.servlet-api-3.1.0.jar:/usr/lib/tez/lib/jersey-client-1.19.jar:/usr/lib/tez/lib/jersey-json-1.19.jar:/usr/lib/tez/lib/jettison-1.3.4.jar:/usr/lib/tez/lib/jsr305-3.0.0.jar:/usr/lib/tez/lib/metrics-core-3.1.0.jar:/usr/lib/tez/lib/netty-all-4.1.72.Final.jar:/usr/lib/tez/lib/protobuf-java-2.5.0.jar:/usr/lib/tez/lib/RoaringBitmap-0.7.45.jar:/usr/lib/tez/lib/slf4j-api-1.7.36.jar:/usr/lib/hadoop-lzo/lib/hadoop-lzo-0.4.19.jar:/usr/lib/hadoop-lzo/lib/hadoop-lzo.jar:/usr/lib/hadoop-lzo/lib/native:/usr/share/aws/aws-java-sdk/aws-java-sdk-bundle-1.12.331.jar:/usr/share/aws/aws-java-sdk/LICENSE.txt:/usr/share/aws/aws-java-sdk/NOTICE.txt:/usr/share/aws/aws-java-sdk/README.md:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/animal-sniffer-annotations-1.14.jar:/usr/share/aws/emr/emrfs/lib/annotations-16.0.2.jar:/usr/share/aws/emr/emrfs/lib/aopalliance-1.0.jar:/usr/share/aws/emr/emrfs/lib/bcprov-ext-jdk15on-1.66.jar:/usr/share/aws/emr/emrfs/lib/checker-qual-2.5.2.jar:/usr/share/aws/emr/emrfs/lib/emrfs-hadoop-assembly-2.54.0.jar:/usr/share/aws/emr/emrfs/lib/error_prone_annotations-2.1.3.jar:/usr/share/aws/emr/emrfs/lib/findbugs-annotations-3.0.1.jar:/usr/share/aws/emr/emrfs/lib/ion-java-1.0.2.jar:/usr/share/aws/emr/emrfs/lib/j2objc-annotations-1.1.jar:/usr/share/aws/emr/emrfs/lib/javax.inject-1.jar:/usr/share/aws/emr/emrfs/lib/jmespath-java-1.12.331.jar:/usr/share/aws/emr/emrfs/lib/jsr305-3.0.2.jar:/usr/share/aws/emr/emrfs/auxlib/:/usr/share/aws/emr/ddb/lib/emr-ddb-hadoop.jar:/usr/share/aws/emr/goodies/lib/emr-hadoop-goodies.jar:/usr/share/aws/emr/kinesis/lib/emr-kinesis-hadoop.jar:/usr/share/aws/emr/cloudwatch-sink/lib/cloudwatch-sink-2.3.0.jar:/usr/share/aws/emr/cloudwatch-sink/lib/cloudwatch-sink.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/usr/lib/flink/lib/log4j-slf4j-impl-2.17.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/usr/lib/hadoop/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> 2022-12-07 03:17:05,554 INFO  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>              [] - Found Yarn properties file under 
> /var/lib/flink/yarn/.yarn-properties-hadoop.
> 2022-12-07 03:17:05,554 INFO  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>              [] - Found Yarn properties file under 
> /var/lib/flink/yarn/.yarn-properties-hadoop.
>                                    ????????
>                                ????????????????
>                             ???????        ???????  ?
>                           ????   ?????????      ?????
>                           ???         ???????    ?????
>                             ???            ???   ?????
>                               ??       ???????????????
>                             ?? ?   ???       ?????? ?????
>                             ?????   ????      ????? ?????
>                          ???????       ???    ??????? ???
>                    ????????? ??         ??    ??????????
>                   ????????  ??           ?   ?? ???????
>                 ????  ???            ?  ?? ???????? ?????
>                ???? ? ??          ? ?? ????????    ????  ??
>               ???? ????          ??????????       ??? ?? ????
>            ???? ?? ???       ???????????         ????  ? ?  ???
>            ???  ?? ??? ?????????              ????           ???
>            ??    ? ???????              ????????          ??? ??
>            ???    ???    ????????????????????            ????  ?
>           ????? ???   ??????   ????????                  ????  ??
>           ????????  ???????????????                            ??
>           ?? ????   ???????  ???       ??????    ??          ???
>           ??? ???  ???  ???????            ????   ?????????????
>            ??? ?????  ????  ??                ??      ????   ???
>            ??   ???   ?     ??                ??              ??
>             ??   ??         ??                 ??        ????????
>              ?? ?????       ??                  ???????????    ??
>               ??   ????      ?                    ???????      ??
>                ???   ?????                         ?? ???????????
>                 ????    ????                     ??????? ????????
>                   ?????                          ??  ????  ?????
>                       ?????????????????????????????????  ?????
>     ______ _ _       _       _____  ____  _         _____ _ _            _  
> BETA
>    |  ___| ()     | |     / ___|/ __ | |       / ____| ()          | |
>    | |__  | |_ _ __ | | __ | (___ | |  | | |      | |    | |_  ___ _ __ | |_
>    |  _| | | | ' | |/ /  ___ | |  | | |      | |    | | |/ _ \ '_ | __|
>    | |    | | | | | |   <   ___) | || | |___  | |___| | |  / | | | |
>    ||    |||| |||_\ |____/ ______|  ___|||_|| ||_|
>         Welcome! Enter 'HELP;' to list all available commands. 'QUIT;' to 
> exit.
> Command history file path: /home/hadoop/.flink-sql-history
> Flink SQL> show jars;
> Empty set
> {code}
> This works only when keeping the hudi-flink bundle jar in /flink/lib/ path 
>  
> But this works with flink 1.15.2 
> {code:java}
> /usr/lib/flink/bin/sql-client.sh embedded -j hudi-flink-bundle_2.12-0.10.1.jar
> Setting HBASE_CONF_DIR=/etc/hbase/conf because no HBASE_CONF_DIR was set.
> Using the result of 'hadoop classpath' to augment the Hadoop classpath: 
> /etc/hadoop/conf:/usr/lib/hadoop/lib/:/usr/lib/hadoop/.//:/usr/lib/hadoop-hdfs/./:/usr/lib/hadoop-hdfs/lib/:/usr/lib/hadoop-hdfs/.//:/usr/lib/hadoop-mapreduce/.//:/usr/lib/hadoop-yarn/lib/:/usr/lib/hadoop-yarn/.//:/etc/tez/conf:/usr/lib/tez/hadoop-shim-0.10.2-amzn-0.jar:/usr/lib/tez/hadoop-shim-2.8-0.10.2-amzn-0.jar:/usr/lib/tez/lib:/usr/lib/tez/LICENSE:/usr/lib/tez/LICENSE-BSD-3clause:/usr/lib/tez/LICENSE-CDDLv1.1-GPLv2_withCPE:/usr/lib/tez/LICENSE-MIT:/usr/lib/tez/LICENSE-SIL_OpenFontLicense-v1.1:/usr/lib/tez/NOTICE:/usr/lib/tez/tez-api-0.10.2-amzn-0.jar:/usr/lib/tez/tez-aux-services-0.10.2-amzn-0.jar:/usr/lib/tez/tez-build-tools-0.10.2-amzn-0.jar:/usr/lib/tez/tez-common-0.10.2-amzn-0.jar:/usr/lib/tez/tez-dag-0.10.2-amzn-0.jar:/usr/lib/tez/tez-examples-0.10.2-amzn-0.jar:/usr/lib/tez/tez-history-parser-0.10.2-amzn-0.jar:/usr/lib/tez/tez-javadoc-tools-0.10.2-amzn-0.jar:/usr/lib/tez/tez-job-analyzer-0.10.2-amzn-0.jar:/usr/lib/tez/tez-mapreduce-0.10.2-amzn-0.jar:/usr/lib/tez/tez-protobuf-history-plugin-0.10.2-amzn-0.jar:/usr/lib/tez/tez-runtime-internals-0.10.2-amzn-0.jar:/usr/lib/tez/tez-runtime-library-0.10.2-amzn-0.jar:/usr/lib/tez/tez-tests-0.10.2-amzn-0.jar:/usr/lib/tez/tez-ui-0.10.2-amzn-0.war:/usr/lib/tez/tez-yarn-timeline-cache-plugin-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-with-acls-0.10.2-amzn-0.jar:/usr/lib/tez/tez-yarn-timeline-history-with-fs-0.10.2-amzn-0.jar:/usr/lib/tez/lib/async-http-client-2.12.3.jar:/usr/lib/tez/lib/commons-cli-1.2.jar:/usr/lib/tez/lib/commons-codec-1.11.jar:/usr/lib/tez/lib/commons-collections4-4.1.jar:/usr/lib/tez/lib/commons-io-2.8.0.jar:/usr/lib/tez/lib/commons-lang-2.6.jar:/usr/lib/tez/lib/guava-31.1-jre.jar:/usr/lib/tez/lib/hadoop-annotations.jar:/usr/lib/tez/lib/hadoop-auth.jar:/usr/lib/tez/lib/hadoop-hdfs-client-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-common-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-common.jar:/usr/lib/tez/lib/hadoop-mapreduce-client-core-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-yarn-server-timeline-pluginstorage-3.3.3-amzn-0.jar:/usr/lib/tez/lib/hadoop-yarn-server-web-proxy.jar:/usr/lib/tez/lib/hhadoop-mapreduce-client-core.jar:/usr/lib/tez/lib/javax.servlet-api-3.1.0.jar:/usr/lib/tez/lib/jersey-client-1.19.jar:/usr/lib/tez/lib/jersey-json-1.19.jar:/usr/lib/tez/lib/jettison-1.3.4.jar:/usr/lib/tez/lib/jsr305-3.0.0.jar:/usr/lib/tez/lib/metrics-core-3.1.0.jar:/usr/lib/tez/lib/netty-all-4.1.72.Final.jar:/usr/lib/tez/lib/protobuf-java-2.5.0.jar:/usr/lib/tez/lib/RoaringBitmap-0.7.45.jar:/usr/lib/tez/lib/slf4j-api-1.7.36.jar:/usr/lib/hadoop-lzo/lib/hadoop-lzo-0.4.19.jar:/usr/lib/hadoop-lzo/lib/hadoop-lzo.jar:/usr/lib/hadoop-lzo/lib/native:/usr/share/aws/aws-java-sdk/aws-java-sdk-bundle-1.12.331.jar:/usr/share/aws/aws-java-sdk/LICENSE.txt:/usr/share/aws/aws-java-sdk/NOTICE.txt:/usr/share/aws/aws-java-sdk/README.md:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/animal-sniffer-annotations-1.14.jar:/usr/share/aws/emr/emrfs/lib/annotations-16.0.2.jar:/usr/share/aws/emr/emrfs/lib/aopalliance-1.0.jar:/usr/share/aws/emr/emrfs/lib/bcprov-ext-jdk15on-1.66.jar:/usr/share/aws/emr/emrfs/lib/checker-qual-2.5.2.jar:/usr/share/aws/emr/emrfs/lib/emrfs-hadoop-assembly-2.54.0.jar:/usr/share/aws/emr/emrfs/lib/error_prone_annotations-2.1.3.jar:/usr/share/aws/emr/emrfs/lib/findbugs-annotations-3.0.1.jar:/usr/share/aws/emr/emrfs/lib/ion-java-1.0.2.jar:/usr/share/aws/emr/emrfs/lib/j2objc-annotations-1.1.jar:/usr/share/aws/emr/emrfs/lib/javax.inject-1.jar:/usr/share/aws/emr/emrfs/lib/jmespath-java-1.12.331.jar:/usr/share/aws/emr/emrfs/lib/jsr305-3.0.2.jar:/usr/share/aws/emr/emrfs/auxlib/:/usr/share/aws/emr/ddb/lib/emr-ddb-hadoop.jar:/usr/share/aws/emr/goodies/lib/emr-hadoop-goodies.jar:/usr/share/aws/emr/kinesis/lib/emr-kinesis-hadoop.jar:/usr/share/aws/emr/cloudwatch-sink/lib/cloudwatch-sink-2.3.0.jar:/usr/share/aws/emr/cloudwatch-sink/lib/cloudwatch-sink.jar
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/usr/lib/flink/lib/log4j-slf4j-impl-2.17.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/usr/lib/hadoop/lib/slf4j-reload4j-1.7.36.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> 2022-12-07 03:19:37,344 INFO  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>              [] - Found Yarn properties file under 
> /var/lib/flink/yarn/.yarn-properties-hadoop.
> 2022-12-07 03:19:37,344 INFO  org.apache.flink.yarn.cli.FlinkYarnSessionCli   
>              [] - Found Yarn properties file under 
> /var/lib/flink/yarn/.yarn-properties-hadoop.
>                                    ????????
>                                ????????????????
>                             ???????        ???????  ?
>                           ????   ?????????      ?????
>                           ???         ???????    ?????
>                             ???            ???   ?????
>                               ??       ???????????????
>                             ?? ?   ???       ?????? ?????
>                             ?????   ????      ????? ?????
>                          ???????       ???    ??????? ???
>                    ????????? ??         ??    ??????????
>                   ????????  ??           ?   ?? ???????
>                 ????  ???            ?  ?? ???????? ?????
>                ???? ? ??          ? ?? ????????    ????  ??
>               ???? ????          ??????????       ??? ?? ????
>            ???? ?? ???       ???????????         ????  ? ?  ???
>            ???  ?? ??? ?????????              ????           ???
>            ??    ? ???????              ????????          ??? ??
>            ???    ???    ????????????????????            ????  ?
>           ????? ???   ??????   ????????                  ????  ??
>           ????????  ???????????????                            ??
>           ?? ????   ???????  ???       ??????    ??          ???
>           ??? ???  ???  ???????            ????   ?????????????
>            ??? ?????  ????  ??                ??      ????   ???
>            ??   ???   ?     ??                ??              ??
>             ??   ??         ??                 ??        ????????
>              ?? ?????       ??                  ???????????    ??
>               ??   ????      ?                    ???????      ??
>                ???   ?????                         ?? ???????????
>                 ????    ????                     ??????? ????????
>                   ?????                          ??  ????  ?????
>                       ?????????????????????????????????  ?????
>     ______ _ _       _       _____  ____  _         _____ _ _            _  
> BETA
>    |  ___| ()     | |     / ___|/ __ | |       / ____| ()          | |
>    | |__  | |_ _ __ | | __ | (___ | |  | | |      | |    | |_  ___ _ __ | |_
>    |  _| | | | ' | |/ /  ___ | |  | | |      | |    | | |/ _ \ '_ | __|
>    | |    | | | | | |   <   ___) | || | |___  | |___| | |  / | | | |
>    ||    |||| |||_\ |____/ ______|  ___|||_|| ||_|
>         Welcome! Enter 'HELP;' to list all available commands. 'QUIT;' to 
> exit.
> Command history file path: /home/hadoop/.flink-sql-history
> Flink SQL> show jars;
> /usr/lib/flink/opt/flink-sql-client-1.15.2.jar
> /usr/lib/flink/opt/flink-python_2.12-1.15.2.jar
> /home/hadoop/hudi-flink-bundle_2.12-0.10.1.jar
> {code}
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to