[ https://issues.apache.org/jira/browse/FLINK-15238?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
xiaojin.wy closed FLINK-15238. ------------------------------ Resolution: Invalid > A sql can't generate a valid execution plan > ------------------------------------------- > > Key: FLINK-15238 > URL: https://issues.apache.org/jira/browse/FLINK-15238 > Project: Flink > Issue Type: Bug > Components: Table SQL / Planner > Affects Versions: 1.10.0 > Reporter: xiaojin.wy > Priority: Major > > *The table and the query is like this:* > CREATE TABLE `scott_emp` ( > > empno INT, > > ename VARCHAR, > > job VARCHAR, > > mgr INT, > > hiredate DATE, > > sal DOUBLE, > > comm DOUBLE, > > deptno INT > > ) WITH ( > > 'format.field-delimiter'='|', > > 'connector.type'='filesystem', > > 'format.derive-schema'='true', > > 'connector.path'='/defender_test_data/daily_regression_blink_sql_1.10/test_scalar/sources/scott_emp.csv', > > 'format.type'='csv' > > ); > CREATE TABLE `scott_dept` ( > > deptno INT, > > dname VARCHAR, > > loc VARCHAR > > ) WITH ( > > 'format.field-delimiter'='|', > > 'connector.type'='filesystem', > > 'format.derive-schema'='true', > > 'connector.path'='/defender_test_data/daily_regression_blink_sql_1.10/test_scalar/sources/scott_dept.csv', > > 'format.type'='csv' > > ); > select deptno, (select empno from scott_emp order by empno limit 1) as x from > scott_dept; > > > *After execution the sql, the exception will appear:* > [ERROR] Could not execute SQL statement. Reason: > org.apache.flink.table.api.TableException: Cannot generate a valid execution > plan for the given query: > LogicalProject(deptno=[$0], x=[$3]) > LogicalJoin(condition=[true], joinType=[left]) > LogicalTableScan(table=[[default_catalog, default_database, scott_dept]]) > LogicalSort(sort0=[$0], dir0=[ASC], fetch=[1]) > LogicalProject(empno=[$0]) > LogicalTableScan(table=[[default_catalog, default_database, scott_emp]]) > This exception indicates that the query uses an unsupported SQL feature. > Please check the documentation for the set of currently supported SQL > features. > > > *The whole exception is:* > Caused by: org.apache.flink.table.api.TableException: Cannot generate a valid > execution plan for the given query:Caused by: > org.apache.flink.table.api.TableException: Cannot generate a valid execution > plan for the given query: > LogicalProject(deptno=[$0], x=[$3]) LogicalJoin(condition=[true], > joinType=[left]) LogicalTableScan(table=[[default_catalog, > default_database, scott_dept]]) LogicalSort(sort0=[$0], dir0=[ASC], > fetch=[1]) LogicalProject(empno=[$0]) > LogicalTableScan(table=[[default_catalog, default_database, scott_emp]]) > This exception indicates that the query uses an unsupported SQL > feature.Please check the documentation for the set of currently supported SQL > features. at > org.apache.flink.table.plan.Optimizer.runVolcanoPlanner(Optimizer.scala:284) > at > org.apache.flink.table.plan.Optimizer.optimizeLogicalPlan(Optimizer.scala:199) > at > org.apache.flink.table.plan.StreamOptimizer.optimize(StreamOptimizer.scala:66) > at > org.apache.flink.table.planner.StreamPlanner.translateToType(StreamPlanner.scala:389) > at > org.apache.flink.table.planner.StreamPlanner.writeToRetractSink(StreamPlanner.scala:308) > at > org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$writeToSink(StreamPlanner.scala:272) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:166) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$2.apply(StreamPlanner.scala:145) > at scala.Option.map(Option.scala:146) at > org.apache.flink.table.planner.StreamPlanner.org$apache$flink$table$planner$StreamPlanner$$translate(StreamPlanner.scala:145) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117) > at > org.apache.flink.table.planner.StreamPlanner$$anonfun$translate$1.apply(StreamPlanner.scala:117) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at > scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) > at scala.collection.Iterator$class.foreach(Iterator.scala:891) at > scala.collection.AbstractIterator.foreach(Iterator.scala:1334) at > scala.collection.IterableLike$class.foreach(IterableLike.scala:72) at > scala.collection.AbstractIterable.foreach(Iterable.scala:54) at > scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at > scala.collection.AbstractTraversable.map(Traversable.scala:104) at > org.apache.flink.table.planner.StreamPlanner.translate(StreamPlanner.scala:117) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:680) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.insertIntoInternal(TableEnvironmentImpl.java:353) > at > org.apache.flink.table.api.internal.TableEnvironmentImpl.insertInto(TableEnvironmentImpl.java:341) > at > org.apache.flink.table.api.internal.TableImpl.insertInto(TableImpl.java:428) > at > org.apache.flink.table.client.gateway.local.LocalExecutor.lambda$executeQueryInternal$12(LocalExecutor.java:640) > at > org.apache.flink.table.client.gateway.local.ExecutionContext.wrapClassLoader(ExecutionContext.java:227) > at > org.apache.flink.table.client.gateway.local.LocalExecutor.executeQueryInternal(LocalExecutor.java:638) > ... 8 more -- This message was sent by Atlassian Jira (v8.3.4#803005)