This is an automated email from the ASF dual-hosted git repository. jark pushed a commit to branch release-1.14 in repository https://gitbox.apache.org/repos/asf/flink.git
The following commit(s) were added to refs/heads/release-1.14 by this push: new 08dbb6d5eff [FLINK-30679][connectors/hive] Fix IndexOutOfBoundsException for Hive lookup join when column pushdown to Hive lookup table source (#21782) 08dbb6d5eff is described below commit 08dbb6d5eff46e72af412643f80ac353636821de Author: hehuiyuan <471627...@qq.com> AuthorDate: Tue Jan 31 11:07:32 2023 +0800 [FLINK-30679][connectors/hive] Fix IndexOutOfBoundsException for Hive lookup join when column pushdown to Hive lookup table source (#21782) Co-authored-by: hehuiyuan1 <hehuiy...@jd.com> --- .../org/apache/flink/connectors/hive/HiveLookupTableSource.java | 4 ++-- .../org/apache/flink/connectors/hive/HiveLookupJoinITCase.java | 8 ++++---- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/HiveLookupTableSource.java b/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/HiveLookupTableSource.java index a97eec16ca2..e5a249b43c5 100644 --- a/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/HiveLookupTableSource.java +++ b/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/HiveLookupTableSource.java @@ -257,8 +257,8 @@ public class HiveLookupTableSource extends HiveTableSource implements LookupTabl jobConf, hiveVersion, tablePath, - getProducedTableSchema().getFieldDataTypes(), - getProducedTableSchema().getFieldNames(), + getTableSchema().getFieldDataTypes(), + getTableSchema().getFieldNames(), catalogTable.getPartitionKeys(), projectedFields, flinkConf.get(HiveOptions.TABLE_EXEC_HIVE_FALLBACK_MAPRED_READER)); diff --git a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java index 6756559755a..f5e3c3f4b42 100644 --- a/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java +++ b/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveLookupJoinITCase.java @@ -365,17 +365,17 @@ public class HiveLookupJoinITCase { batchEnv.registerCatalog(hiveCatalog.getName(), hiveCatalog); batchEnv.useCatalog(hiveCatalog.getName()); batchEnv.executeSql( - "insert overwrite bounded_table values (1,'a',10),(2,'a',21),(2,'b',22),(3,'c',33)") + "insert overwrite bounded_table values (1,'a',10),(2,'b',22),(3,'c',33)") .await(); tableEnv.getConfig().setSqlDialect(SqlDialect.DEFAULT); TableImpl flinkTable = (TableImpl) tableEnv.sqlQuery( - "select b.x, b.y from " + "select b.x, b.z from " + " default_catalog.default_database.probe as p " - + " join bounded_table for system_time as of p.p as b on p.x=b.x and p.y=b.y"); + + " join bounded_table for system_time as of p.p as b on p.x=b.x"); List<Row> results = CollectionUtil.iteratorToList(flinkTable.execute().collect()); - assertEquals("[+I[1, a], +I[2, b], +I[3, c]]", results.toString()); + assertEquals("[+I[1, 10], +I[1, 10], +I[2, 22], +I[2, 22], +I[3, 33]]", results.toString()); } @Test