[ 
https://issues.apache.org/jira/browse/FLINK-8492?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16336105#comment-16336105
 ] 

Fabian Hueske commented on FLINK-8492:
--------------------------------------

OK, I think I found a simple fix. The problem is in fact a cost modeling issue.
We do not penalize having multiple calcs due to the cost function in 
{{CommonCalc.computeSelfCost()}}.

If we add a constant {{1}} to {{compCnt}}, the calcs will be merged because 
this reduces the cost.

{code}
private[flink] def computeSelfCost(
      calcProgram: RexProgram,
      planner: RelOptPlanner,
      rowCnt: Double): RelOptCost = {

    // compute number of expressions that do not access a field or literal, 
i.e. computations,
    // conditions, etc. We only want to account for computations, not for 
simple projections.
    // CASTs in RexProgram are reduced as far as possible by 
ReduceExpressionsRule
    // in normalization stage. So we should ignore CASTs here in optimization 
stage.
    val compCnt = calcProgram.getExprList.asScala.toList.count {
      case _: RexInputRef => false
      case _: RexLiteral => false
      case c: RexCall if c.getOperator.getName.equals("CAST") => false
      case _ => true
    } + 1   // <------ THIS IS THE FIX

    val newRowCnt = estimateRowCount(calcProgram, rowCnt)
    planner.getCostFactory.makeCost(newRowCnt, newRowCnt * compCnt, 0)
  }
{code}


> Fix unsupported exception for udtf with multi calc
> --------------------------------------------------
>
>                 Key: FLINK-8492
>                 URL: https://issues.apache.org/jira/browse/FLINK-8492
>             Project: Flink
>          Issue Type: Bug
>          Components: Table API &amp; SQL
>            Reporter: Hequn Cheng
>            Assignee: Hequn Cheng
>            Priority: Major
>
> Considering the following test, unsupported exception will be thrown due to 
> multi calc existing between correlate and TableFunctionScan.
> {code:java}
> // code placeholder
> @Test
> def testCrossJoinWithMultiFilter(): Unit = {
>   val t = testData(env).toTable(tEnv).as('a, 'b, 'c)
>   val func0 = new TableFunc0
>   val result = t
>     .join(func0('c) as('d, 'e))
>     .select('c, 'd, 'e)
>     .where('e > 10)
>     .where('e > 20)
>     .select('c, 'd)
>     .toAppendStream[Row]
>   result.addSink(new StreamITCase.StringSink[Row])
>   env.execute()
>   val expected = mutable.MutableList("Jack#22,Jack,22", "Anna#44,Anna,44")
>   assertEquals(expected.sorted, StreamITCase.testResults.sorted)
> }
> {code}
> I can see two options to fix this problem:
>  # Adapt calcite OptRule to merge the continuous calc.
>  # Merge multi calc in correlate convert rule.
> I prefer the second one, not only it is easy to implement but also i think 
> with or without an optimize rule should not influence flink functionality. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to