[GitHub] [flink] flinkbot edited a comment on pull request #15004: [FLINK-21253][table-planner-blink] Support grouping set syntax for GroupWindowAggregate

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15004:
URL: https://github.com/apache/flink/pull/15004#issuecomment-785052726


   
   ## CI report:
   
   * 3af09855fc47130d93b11e6d6ba1c3dedb6574a5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=13733)
 
   * e2d2a09c2abc55337de23906b4aa348d3502f50d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-21627) When you insert multiple inserts with statementSet, you modify multiple inserts with OPTIONS('table-name '=' XXX '), but only the first one takes effect

2021-04-17 Thread godfrey he (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

godfrey he closed FLINK-21627.
--
Resolution: Fixed

Fixed in 1.13.0: c6147f2df2e81c1b1db6aa43b3a827a8c4148af1

> When you insert multiple inserts with statementSet, you modify multiple 
> inserts with OPTIONS('table-name '=' XXX '), but only the first one takes 
> effect
> 
>
> Key: FLINK-21627
> URL: https://issues.apache.org/jira/browse/FLINK-21627
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.12.0
>Reporter: wangtaiyang
>Assignee: godfrey he
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> {code:java}
> //代码占位符
> StatementSet statementSet = tableEnvironment.createStatementSet();
> String sql1 = "insert into test select a,b,c from test_a_12342 /*+
> OPTIONS('table-name'='test_a_1')*/";
> String sql2 = "insert into test select a,b,c from test_a_12342 /*+
> OPTIONS('table-name'='test_a_2')*/";
> statementSet.addInsertSql(sql1);
> statementSet.addInsertSql(sql2);
> statementSet.execute();
> {code}
> Sql code as above, in the final after the insert is put test_a_1 table data 
> into the two times, and test_a_2 data did not insert, is excuse me this bug



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #14830: [FLINK-20487][table-planner-blink] Support to consume retractions for window aggregate operator

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #14830:
URL: https://github.com/apache/flink/pull/14830#issuecomment-770878871


   
   ## CI report:
   
   * a2587cacaa7abea0059a6aadf16cbb1c5a4f31b3 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=13700)
 
   * c39b7a4ecde12aca916b2228ffe734598749fa9d Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16721)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe closed pull request #15559: [FLINK-21627][table-planner-blink] The digest of TableScan & Sink should consider table hints

2021-04-17 Thread GitBox


godfreyhe closed pull request #15559:
URL: https://github.com/apache/flink/pull/15559


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14830: [FLINK-20487][table-planner-blink] Support to consume retractions for window aggregate operator

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #14830:
URL: https://github.com/apache/flink/pull/14830#issuecomment-770878871


   
   ## CI report:
   
   * a2587cacaa7abea0059a6aadf16cbb1c5a4f31b3 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=13700)
 
   * c39b7a4ecde12aca916b2228ffe734598749fa9d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-20487) Support to consume retractions for window aggregate operator

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324415#comment-17324415
 ] 

Andy commented on FLINK-20487:
--

[~flink-jira-bot] [~jark] [~godfreyhe] I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.

> Support to consume retractions for window aggregate operator
> 
>
> Key: FLINK-20487
> URL: https://issues.apache.org/jira/browse/FLINK-20487
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Planner
>Reporter: jiayue.yu
>Assignee: Andy
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.14.0
>
>
> {code}
> EXCEPTION: org.apache.flink.table.api.TableException: Group Window Aggregate: 
> Retraction on windowed GroupBy Aggregate is not supported yet. 
> org.apache.flink.table.api.TableException: Group Window Aggregate: Retraction 
> on windowed GroupBy Aggregate is not supported yet. please re-check sql 
> grammar. Note: Windowed GroupBy Aggregate should not follow anon-windowed 
> GroupBy aggregation. at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecGroupWindowAggregateBase.translateToPlanInternal(StreamExecGroupWindowAggregateBase.scala:138)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecGroupWindowAggregateBase.translateToPlanInternal(StreamExecGroupWindowAggregateBase.scala:54)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan(ExecNode.scala:58)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan$(ExecNode.scala:56)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecGroupWindowAggregateBase.translateToPlan(StreamExecGroupWindowAggregateBase.scala:54)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecCalc.translateToPlanInternal(StreamExecCalc.scala:54)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecCalc.translateToPlanInternal(StreamExecCalc.scala:39)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan(ExecNode.scala:58)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan$(ExecNode.scala:56)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecCalcBase.translateToPlan(StreamExecCalcBase.scala:38)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToTransformation(StreamExecSink.scala:184)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:91)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:48)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan(ExecNode.scala:58)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan$(ExecNode.scala:56)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlan(StreamExecSink.scala:48)
>  at 
> org.apache.flink.table.planner.delegation.StreamPlanner.$anonfun$translateToPlan$1(StreamPlanner.scala:60)
>  at 
> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:233) at 
> scala.collection.Iterator.foreach(Iterator.scala:937) at 
> scala.collection.Iterator.foreach$(Iterator.scala:937) at 
> scala.collection.AbstractIterator.foreach(Iterator.scala:1425) at 
> scala.collection.IterableLike.foreach(IterableLike.scala:70) at 
> scala.collection.IterableLike.foreach$(IterableLike.scala:69) at 
> scala.collection.AbstractIterable.foreach(Iterable.scala:54) at 
> scala.collection.TraversableLike.map(TraversableLike.scala:233) at 
> scala.collection.TraversableLike.map$(TraversableLike.scala:226) at 
> scala.collection.AbstractTraversable.map(Traversable.scala:104) at 
> org.apache.flink.table.planner.delegation.StreamPlanner.translateToPlan(StreamPlanner.scala:59)
>  at 
> org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:153)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.translate(TableEnvironmentImpl.java:682)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.insertIntoInternal(TableEnvironmentImpl.java:355)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.insertInto(TableEnvironmentImpl.java:334)
> {code}
>  
> CASE:
> {code:sql}
> SELECT
>  DATE_FORMAT(tumble_end(ROWTIME ,interval '1' hour),'-MM-dd HH') as 
> stat_time,
>  count(crypto_customer_number) first_phone_num
> FROM (
>  SELECT 
>  ROWTIME,
>  crypto_customer_number,
>  row_number() over(partition by crypto_customer_number order by ROWTIME ) as 
> rn
>  FROM 

[jira] [Commented] (FLINK-22099) Fix bug for semi/anti window join.

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22099?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324414#comment-17324414
 ] 

Andy commented on FLINK-22099:
--

[~flink-jira-bot] [~jark] [~godfreyhe]  I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.

> Fix bug for semi/anti window join.
> --
>
> Key: FLINK-22099
> URL: https://issues.apache.org/jira/browse/FLINK-22099
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-22098) Fix bug for window join: plan is wrong if join condition contain 'IS NOT DISTINCT FROM'

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324413#comment-17324413
 ] 

Andy commented on FLINK-22098:
--

[~flink-jira-bot] [~jark] [~godfreyhe]   I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.
 * [|https://issues.apache.org/jira/secure/AddComment!default.jspa?id=13357018]

> Fix bug for window join: plan is wrong if join condition contain 'IS NOT 
> DISTINCT FROM'
> ---
>
> Key: FLINK-22098
> URL: https://issues.apache.org/jira/browse/FLINK-22098
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Plan test is wrong for window join if join condition contain 'IS NOT DISTINCT 
> FROM'.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-22098) Fix bug for window join: plan is wrong if join condition contain 'IS NOT DISTINCT FROM'

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22098?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324413#comment-17324413
 ] 

Andy edited comment on FLINK-22098 at 4/18/21, 5:19 AM:


[~flink-jira-bot] [~jark] [~godfreyhe]   I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.


was (Author: qingru zhang):
[~flink-jira-bot] [~jark] [~godfreyhe]   I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.
 * [|https://issues.apache.org/jira/secure/AddComment!default.jspa?id=13357018]

> Fix bug for window join: plan is wrong if join condition contain 'IS NOT 
> DISTINCT FROM'
> ---
>
> Key: FLINK-22098
> URL: https://issues.apache.org/jira/browse/FLINK-22098
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Plan test is wrong for window join if join condition contain 'IS NOT DISTINCT 
> FROM'.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #15479: [FLINK-19606][table-runtime-blink] Implement streaming window join operator

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15479:
URL: https://github.com/apache/flink/pull/15479#issuecomment-812344008


   
   ## CI report:
   
   * 93b9b3d8dac2c3ecd13e3e3ede46f4a73263e34d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16015)
 
   * 36663891630b1ca8dbe82c3225286ffab432bb18 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16720)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15477: [FLINK-22099][table-planner-blink] Fix bug for semi/anti window join.

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15477:
URL: https://github.com/apache/flink/pull/15477#issuecomment-812337208


   
   ## CI report:
   
   * 68bacb55c3c718eddba8cf5596bf94c3e4931d40 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16011)
 
   * fb7df56df7e4b5e1cf9f1eee726f9d338c0187e6 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16719)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15476: [FLINK-22098][table-planner-blink] Fix bug for window join: plan is wrong if join condition contain 'IS NOT DISTINCT FROM'

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15476:
URL: https://github.com/apache/flink/pull/15476#issuecomment-812307606


   
   ## CI report:
   
   * 3023dfa33790b3140a132fd2f76f56a2caad4a78 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16009)
 
   * f85197c7bb0c334252f289fd440e041ebb8ca26f Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16718)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15479: [FLINK-19606][table-runtime-blink] Implement streaming window join operator

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15479:
URL: https://github.com/apache/flink/pull/15479#issuecomment-812344008


   
   ## CI report:
   
   * 93b9b3d8dac2c3ecd13e3e3ede46f4a73263e34d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16015)
 
   * 36663891630b1ca8dbe82c3225286ffab432bb18 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15477: [FLINK-22099][table-planner-blink] Fix bug for semi/anti window join.

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15477:
URL: https://github.com/apache/flink/pull/15477#issuecomment-812337208


   
   ## CI report:
   
   * 68bacb55c3c718eddba8cf5596bf94c3e4931d40 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16011)
 
   * fb7df56df7e4b5e1cf9f1eee726f9d338c0187e6 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15476: [FLINK-22098][table-planner-blink] Fix bug for window join: plan is wrong if join condition contain 'IS NOT DISTINCT FROM'

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15476:
URL: https://github.com/apache/flink/pull/15476#issuecomment-812307606


   
   ## CI report:
   
   * 3023dfa33790b3140a132fd2f76f56a2caad4a78 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16009)
 
   * f85197c7bb0c334252f289fd440e041ebb8ca26f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] hehuiyuan removed a comment on pull request #15615: [FLINK-22021][table-planner-blink] Fix exception when PushFilterIntoLegacyTableSourceScanRule is faced with INTERVAL types

2021-04-17 Thread GitBox


hehuiyuan removed a comment on pull request #15615:
URL: https://github.com/apache/flink/pull/15615#issuecomment-821067136


   @flinkbot re-run


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-21302) Fix NPE when use row_number() in over agg

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-21302?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324405#comment-17324405
 ] 

Andy commented on FLINK-21302:
--

[~flink-jira-bot] [~jark] [~godfreyhe]   I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.

> Fix NPE when use row_number() in over agg
> -
>
> Key: FLINK-21302
> URL: https://issues.apache.org/jira/browse/FLINK-21302
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> `NullPointException` would be thrown out if SQL contains row_number() in Over 
> Aggregate.
> {code:scala}
> @Test
> def testRowNumberOnOver(): Unit = {
>   val t = failingDataSource(TestData.tupleData5)
> .toTable(tEnv, 'a, 'b, 'c, 'd, 'e, 'proctime.proctime)
>   tEnv.registerTable("MyTable", t)
>   val sqlQuery = "SELECT a, ROW_NUMBER() OVER (PARTITION BY a ORDER BY 
> proctime()) FROM MyTable"
>   val sink = new TestingAppendSink
>   tEnv.sqlQuery(sqlQuery).toAppendStream[Row].addSink(sink)
>   env.execute()
> }{code}
>  
> The following exception would be thrown out.
> {code:java}
> java.lang.NullPointerExceptionjava.lang.NullPointerException at 
> scala.collection.mutable.ArrayOps$ofInt$.length$extension(ArrayOps.scala:240) 
> at scala.collection.mutable.ArrayOps$ofInt.length(ArrayOps.scala:240) at 
> scala.collection.SeqLike$class.size(SeqLike.scala:106) at 
> scala.collection.mutable.ArrayOps$ofInt.size(ArrayOps.scala:234) at 
> scala.collection.mutable.Builder$class.sizeHint(Builder.scala:69) at 
> scala.collection.mutable.ArrayBuilder.sizeHint(ArrayBuilder.scala:22) at 
> scala.collection.TraversableLike$class.builder$1(TraversableLike.scala:230) 
> at scala.collection.TraversableLike$class.map(TraversableLike.scala:233) at 
> scala.collection.mutable.ArrayOps$ofInt.map(ArrayOps.scala:234) at 
> org.apache.flink.table.planner.codegen.agg.DeclarativeAggCodeGen.(DeclarativeAggCodeGen.scala:82)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator$$anonfun$3.apply(AggsHandlerCodeGenerator.scala:222)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator$$anonfun$3.apply(AggsHandlerCodeGenerator.scala:214)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>  at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at 
> scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at 
> scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186) at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator.initialAggregateInformation(AggsHandlerCodeGenerator.scala:214)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator.generateAggsHandler(AggsHandlerCodeGenerator.scala:325)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.stream.StreamExecOverAggregate.createUnboundedOverProcessFunction(StreamExecOverAggregate.java:262)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.stream.StreamExecOverAggregate.translateToPlanInternal(StreamExecOverAggregate.java:154)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecEdge.translateToPlan(ExecEdge.java:247)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecCalc.translateToPlanInternal(CommonExecCalc.java:65)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecEdge.translateToPlan(ExecEdge.java:247)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLegacySink.translateToTransformation(CommonExecLegacySink.java:167)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLegacySink.translateToPlanInternal(CommonExecLegacySink.java:136)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:67)
>  at 
> org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:66)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 

[jira] [Comment Edited] (FLINK-19606) Implement streaming window join operator

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19606?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324403#comment-17324403
 ] 

Andy edited comment on FLINK-19606 at 4/18/21, 4:13 AM:


[~flink-jira-bot] [~jark] [~godfreyhe]. I'm still working on the issue, please 
take a look at pull request when you have time, thanks a lot.


was (Author: qingru zhang):
[~flink-jira-bot] [~jark] [~godfreyhe]. I'm still working on the issue, please 
take a look when you have time, thanks a lot.

> Implement streaming window join operator
> 
>
> Key: FLINK-19606
> URL: https://issues.apache.org/jira/browse/FLINK-19606
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Runtime
>Reporter: Jark Wu
>Assignee: Andy
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.14.0
>
>
> Implement streaming window join operator in blink runtime.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-21302) Fix NPE when use row_number() in over agg

2021-04-17 Thread Andy (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21302?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andy updated FLINK-21302:
-
Labels: pull-request-available  (was: pull-request-available stale-assigned)

> Fix NPE when use row_number() in over agg
> -
>
> Key: FLINK-21302
> URL: https://issues.apache.org/jira/browse/FLINK-21302
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> `NullPointException` would be thrown out if SQL contains row_number() in Over 
> Aggregate.
> {code:scala}
> @Test
> def testRowNumberOnOver(): Unit = {
>   val t = failingDataSource(TestData.tupleData5)
> .toTable(tEnv, 'a, 'b, 'c, 'd, 'e, 'proctime.proctime)
>   tEnv.registerTable("MyTable", t)
>   val sqlQuery = "SELECT a, ROW_NUMBER() OVER (PARTITION BY a ORDER BY 
> proctime()) FROM MyTable"
>   val sink = new TestingAppendSink
>   tEnv.sqlQuery(sqlQuery).toAppendStream[Row].addSink(sink)
>   env.execute()
> }{code}
>  
> The following exception would be thrown out.
> {code:java}
> java.lang.NullPointerExceptionjava.lang.NullPointerException at 
> scala.collection.mutable.ArrayOps$ofInt$.length$extension(ArrayOps.scala:240) 
> at scala.collection.mutable.ArrayOps$ofInt.length(ArrayOps.scala:240) at 
> scala.collection.SeqLike$class.size(SeqLike.scala:106) at 
> scala.collection.mutable.ArrayOps$ofInt.size(ArrayOps.scala:234) at 
> scala.collection.mutable.Builder$class.sizeHint(Builder.scala:69) at 
> scala.collection.mutable.ArrayBuilder.sizeHint(ArrayBuilder.scala:22) at 
> scala.collection.TraversableLike$class.builder$1(TraversableLike.scala:230) 
> at scala.collection.TraversableLike$class.map(TraversableLike.scala:233) at 
> scala.collection.mutable.ArrayOps$ofInt.map(ArrayOps.scala:234) at 
> org.apache.flink.table.planner.codegen.agg.DeclarativeAggCodeGen.(DeclarativeAggCodeGen.scala:82)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator$$anonfun$3.apply(AggsHandlerCodeGenerator.scala:222)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator$$anonfun$3.apply(AggsHandlerCodeGenerator.scala:214)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>  at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at 
> scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at 
> scala.collection.mutable.ArrayOps$ofRef.map(ArrayOps.scala:186) at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator.initialAggregateInformation(AggsHandlerCodeGenerator.scala:214)
>  at 
> org.apache.flink.table.planner.codegen.agg.AggsHandlerCodeGenerator.generateAggsHandler(AggsHandlerCodeGenerator.scala:325)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.stream.StreamExecOverAggregate.createUnboundedOverProcessFunction(StreamExecOverAggregate.java:262)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.stream.StreamExecOverAggregate.translateToPlanInternal(StreamExecOverAggregate.java:154)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecEdge.translateToPlan(ExecEdge.java:247)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecCalc.translateToPlanInternal(CommonExecCalc.java:65)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecEdge.translateToPlan(ExecEdge.java:247)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLegacySink.translateToTransformation(CommonExecLegacySink.java:167)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.common.CommonExecLegacySink.translateToPlanInternal(CommonExecLegacySink.java:136)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNodeBase.translateToPlan(ExecNodeBase.java:128)
>  at 
> org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:67)
>  at 
> org.apache.flink.table.planner.delegation.StreamPlanner$$anonfun$translateToPlan$1.apply(StreamPlanner.scala:66)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at scala.collection.Iterator$class.foreach(Iterator.scala:891) at 
> scala.collection.AbstractIterator.foreach(Iterator.scala:1334) at 
> 

[jira] [Updated] (FLINK-21923) SplitAggregateRule will be abnormal, when the sum/count and avg in SQL at the same time

2021-04-17 Thread tartarus (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21923?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

tartarus updated FLINK-21923:
-
Labels: pull-request-available  (was: pull-request-available stale-assigned)

> SplitAggregateRule will be abnormal, when the sum/count and avg in SQL at the 
> same time
> ---
>
> Key: FLINK-21923
> URL: https://issues.apache.org/jira/browse/FLINK-21923
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: tartarus
>Assignee: tartarus
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> SplitAggregateRule optimizes one-layer aggregation to two-layer aggregation 
> to improve computing performance under data skew.
> In the partial phase, avg will be translated into count and sum. If count 
> already exists in the original SQL at this time, the engine will remove the 
> duplicate count, and then add Project to calculate and restore the optimized 
> count result value.
> {code:java}
> relBuilder.aggregate(
>   relBuilder.groupKey(fullGroupSet, 
> ImmutableList.of[ImmutableBitSet](fullGroupSet)),
>   newPartialAggCalls)
> relBuilder.peek().asInstanceOf[FlinkLogicalAggregate]
>   .setPartialFinalType(PartialFinalType.PARTIAL)
> {code}
> so `relBuilder.peek()` will return `FlinkLogicalCalc` not 
> `FlinkLogicalAggregate`,
> then will throw exception like 
> {code:java}
> java.lang.ClassCastException: 
> org.apache.flink.table.planner.plan.nodes.logical.FlinkLogicalCalc cannot be 
> cast to 
> org.apache.flink.table.planner.plan.nodes.logical.FlinkLogicalAggregate
>   at 
> org.apache.flink.table.planner.plan.rules.logical.SplitAggregateRule.onMatch(SplitAggregateRule.scala:286)
>   at 
> org.apache.calcite.plan.AbstractRelOptPlanner.fireRule(AbstractRelOptPlanner.java:333)
>   at org.apache.calcite.plan.hep.HepPlanner.applyRule(HepPlanner.java:542)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.applyRules(HepPlanner.java:407)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.executeInstruction(HepPlanner.java:243)
>   at 
> org.apache.calcite.plan.hep.HepInstruction$RuleInstance.execute(HepInstruction.java:127)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.executeProgram(HepPlanner.java:202)
>   at 
> org.apache.calcite.plan.hep.HepPlanner.findBestExp(HepPlanner.java:189)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepProgram.optimize(FlinkHepProgram.scala:69)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkHepRuleSetProgram.optimize(FlinkHepRuleSetProgram.scala:87)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram$$anonfun$optimize$1.apply(FlinkChainedProgram.scala:62)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram$$anonfun$optimize$1.apply(FlinkChainedProgram.scala:58)
>   at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
>   at 
> scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
>   at scala.collection.Iterator$class.foreach(Iterator.scala:891)
>   at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
>   at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>   at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>   at 
> scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
>   at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:104)
>   at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.optimize(FlinkChainedProgram.scala:57)
>   at 
> org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:163)
>   at 
> org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:79)
>   at 
> org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:77)
>   at 
> org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:284)
>   at 
> org.apache.flink.table.planner.utils.TableTestUtilBase.assertPlanEquals(TableTestBase.scala:889)
>   at 
> org.apache.flink.table.planner.utils.TableTestUtilBase.doVerifyPlan(TableTestBase.scala:780)
>   at 
> org.apache.flink.table.planner.utils.TableTestUtilBase.verifyPlan(TableTestBase.scala:283)
>   at 
> org.apache.flink.table.planner.plan.rules.logical.SplitAggregateRuleTest.testAggWithFilterClause2(SplitAggregateRuleTest.scala:205)
> {code}
> We can 

[jira] [Commented] (FLINK-19606) Implement streaming window join operator

2021-04-17 Thread Andy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19606?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324403#comment-17324403
 ] 

Andy commented on FLINK-19606:
--

[~flink-jira-bot] [~jark] [~godfreyhe]. I'm still working on the issue, please 
take a look when you have time, thanks a lot.

> Implement streaming window join operator
> 
>
> Key: FLINK-19606
> URL: https://issues.apache.org/jira/browse/FLINK-19606
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Runtime
>Reporter: Jark Wu
>Assignee: Andy
>Priority: Major
>  Labels: pull-request-available, stale-assigned
> Fix For: 1.14.0
>
>
> Implement streaming window join operator in blink runtime.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-22099) Fix bug for semi/anti window join.

2021-04-17 Thread Andy (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22099?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andy updated FLINK-22099:
-
Labels: pull-request-available  (was: pull-request-available stale-assigned)

> Fix bug for semi/anti window join.
> --
>
> Key: FLINK-22099
> URL: https://issues.apache.org/jira/browse/FLINK-22099
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-22098) Fix bug for window join: plan is wrong if join condition contain 'IS NOT DISTINCT FROM'

2021-04-17 Thread Andy (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andy updated FLINK-22098:
-
Labels: pull-request-available  (was: pull-request-available stale-assigned)

> Fix bug for window join: plan is wrong if join condition contain 'IS NOT 
> DISTINCT FROM'
> ---
>
> Key: FLINK-22098
> URL: https://issues.apache.org/jira/browse/FLINK-22098
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Andy
>Assignee: Andy
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Plan test is wrong for window join if join condition contain 'IS NOT DISTINCT 
> FROM'.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-22310) The result of LogicalWindowJsonDeserializer is incorrect

2021-04-17 Thread godfrey he (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22310?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

godfrey he closed FLINK-22310.
--
Resolution: Fixed

Fixed in 1.13.0: d32ca07743c729e66656f119f057249e03e5c5e2

> The result of LogicalWindowJsonDeserializer is incorrect
> 
>
> Key: FLINK-22310
> URL: https://issues.apache.org/jira/browse/FLINK-22310
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: godfrey he
>Assignee: godfrey he
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> The reason is the wrong argument is given in LogicalWindowJsonDeserializer 
> when creating FieldReferenceExpression instance, see line#137 in 
> LogicalWindowJsonDeserializer. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] godfreyhe closed pull request #15645: [FLINK-22310][table-planner-blink] Fix the incorrect deserialization result of LogicalWindowJsonDeserializer

2021-04-17 Thread GitBox


godfreyhe closed pull request #15645:
URL: https://github.com/apache/flink/pull/15645


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19614) Further optimization of sort-merge based blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao closed FLINK-19614.
---
Resolution: Resolved

> Further optimization of sort-merge based blocking shuffle
> -
>
> Key: FLINK-19614
> URL: https://issues.apache.org/jira/browse/FLINK-19614
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Assignee: Yingjie Cao
>Priority: Major
> Fix For: 1.13.0
>
>
> FLINK-19582 introduces a basic sort-merge based blocking shuffle 
> implementation. We can further optimize it based on the approaches proposed 
> in 
> [https://docs.google.com/document/d/1mpekX6aAHJhBsQ0pS9MxDiFQjHQIuaJH0GAQHh0GlJ0/edit?usp=sharing|https://docs.google.com/document/d/1mpekX6aAHJhBsQ0pS9MxDiFQjHQIuaJH0GAQHh0GlJ0/edit?usp=sharing,].
> This is the umbrella ticket for the optimizations.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-22307) Increase the default value of data writing cache size (not configurable) for sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao closed FLINK-22307.
---
Resolution: Resolved

> Increase the default value of data writing cache size (not configurable) for 
> sort-merge blocking shuffle
> 
>
> Key: FLINK-22307
> URL: https://issues.apache.org/jira/browse/FLINK-22307
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the data writing cache is 8M, which is not enough if data 
> compression is enabled. By increasing the cache size to 16M, the performance 
> of our benchmark job can be increased by about 20%. (We may make it 
> configurable in the future)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-22307) Increase the default value of data writing cache size (not configurable) for sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22307?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324396#comment-17324396
 ] 

Yingjie Cao commented on FLINK-22307:
-

Merged via d29357791315ace0218308f9f96bf65c7c8079d7 on master.

> Increase the default value of data writing cache size (not configurable) for 
> sort-merge blocking shuffle
> 
>
> Key: FLINK-22307
> URL: https://issues.apache.org/jira/browse/FLINK-22307
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the data writing cache is 8M, which is not enough if data 
> compression is enabled. By increasing the cache size to 16M, the performance 
> of our benchmark job can be increased by about 20%. (We may make it 
> configurable in the future)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-22305) Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22305?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324394#comment-17324394
 ] 

Yingjie Cao commented on FLINK-22305:
-

Merged via 77e0478a7879bd41f9f52872d84c467d158b8098 on master.

> Improve log messages of sort-merge blocking shuffle
> ---
>
> Key: FLINK-22305
> URL: https://issues.apache.org/jira/browse/FLINK-22305
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the default value of taskmanager.network.sort-shuffle.min-buffers 
> is 64, which is pretty small. As suggested, we'd like to increase the default 
> value of taskmanager.network.sort-shuffle.min-buffers. By increasing the 
> default taskmanager.network.sort-shuffle.min-buffers, the corner case of very 
> small in-memory sort-buffer and write-buffer can be avoid, which is better 
> for performance.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-22305) Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22305?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao closed FLINK-22305.
---
Resolution: Resolved

> Improve log messages of sort-merge blocking shuffle
> ---
>
> Key: FLINK-22305
> URL: https://issues.apache.org/jira/browse/FLINK-22305
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the default value of taskmanager.network.sort-shuffle.min-buffers 
> is 64, which is pretty small. As suggested, we'd like to increase the default 
> value of taskmanager.network.sort-shuffle.min-buffers. By increasing the 
> default taskmanager.network.sort-shuffle.min-buffers, the corner case of very 
> small in-memory sort-buffer and write-buffer can be avoid, which is better 
> for performance.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] guoweiM closed pull request #15651: [FLINK-22307][network] Increase the default value of data writing cache size (not configurable) for sort-merge blocking shuffle

2021-04-17 Thread GitBox


guoweiM closed pull request #15651:
URL: https://github.com/apache/flink/pull/15651


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15651: [FLINK-22307][network] Increase the default value of data writing cache size (not configurable) for sort-merge blocking shuffle

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15651:
URL: https://github.com/apache/flink/pull/15651#issuecomment-821767454


   
   ## CI report:
   
   * 3d65a719c16d9aaf8a44f843e2e21d1719947559 UNKNOWN
   * c840d852f79ed239c164c4809f36529bed7ffe5d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16700)
 
   * b955af657e6ef1801b4228c7d07dd63dff9f61ae UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] guoweiM closed pull request #15652: [FLINK-22305][network] Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread GitBox


guoweiM closed pull request #15652:
URL: https://github.com/apache/flink/pull/15652


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] guoweiM commented on a change in pull request #15652: [FLINK-22305][network] Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread GitBox


guoweiM commented on a change in pull request #15652:
URL: https://github.com/apache/flink/pull/15652#discussion_r615332364



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/SortMergeResultPartition.java
##
@@ -160,18 +160,12 @@ public void setup() throws IOException {
 }
 
 int numRequiredBuffer = bufferPool.getNumberOfRequiredMemorySegments();
-String errorMessage =
-String.format(
-"Too few sort buffers, please increase %s to a larger 
value (more than %d).",
-
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS,
-2 * expectedWriteBuffers);
-if (numRequiredBuffer < 2 * expectedWriteBuffers) {
-LOG.warn(errorMessage);
-}
-
 int numWriteBuffers = Math.min(numRequiredBuffer / 2, 
expectedWriteBuffers);
 if (numWriteBuffers < 1) {
-throw new IOException(errorMessage);
+throw new IOException(
+String.format(
+"Too few sort buffers, please increase %s.",
+
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS));

Review comment:
   Thanks for explaining.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Comment Edited] (FLINK-17510) StreamingKafkaITCase. testKafka timeouts on downloading Kafka

2021-04-17 Thread Guowei Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324387#comment-17324387
 ] 

Guowei Ma edited comment on FLINK-17510 at 4/18/21, 2:52 AM:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=2b7514ee-e706-5046-657b-3430666e7bd9=17137

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=68a897ab-3047-5660-245a-cce8f83859f6=d47e27f5-9721-5d5f-1cf3-62adbf3d115d=17334

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16692=logs=68a897ab-3047-5660-245a-cce8f83859f6=d47e27f5-9721-5d5f-1cf3-62adbf3d115d=17405


was (Author: maguowei):
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=2b7514ee-e706-5046-657b-3430666e7bd9=17137

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=68a897ab-3047-5660-245a-cce8f83859f6=d47e27f5-9721-5d5f-1cf3-62adbf3d115d=17334

> StreamingKafkaITCase. testKafka timeouts on downloading Kafka
> -
>
> Key: FLINK-17510
> URL: https://issues.apache.org/jira/browse/FLINK-17510
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Connectors / Kafka, Tests
>Affects Versions: 1.11.3, 1.12.1, 1.13.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=585=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-05-05T00:06:49.7268716Z [INFO] 
> ---
> 2020-05-05T00:06:49.7268938Z [INFO]  T E S T S
> 2020-05-05T00:06:49.7269282Z [INFO] 
> ---
> 2020-05-05T00:06:50.5336315Z [INFO] Running 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8603439Z [ERROR] Tests run: 3, Failures: 0, Errors: 2, 
> Skipped: 0, Time elapsed: 276.323 s <<< FAILURE! - in 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8604882Z [ERROR] testKafka[1: 
> kafka-version:0.11.0.2](org.apache.flink.tests.util.kafka.StreamingKafkaITCase)
>   Time elapsed: 120.024 s  <<< ERROR!
> 2020-05-05T00:11:26.8605942Z java.io.IOException: Process ([wget, -q, -P, 
> /tmp/junit2815750531595874769/downloads/1290570732, 
> https://archive.apache.org/dist/kafka/0.11.0.2/kafka_2.11-0.11.0.2.tgz]) 
> exceeded timeout (12) or number of retries (3).
> 2020-05-05T00:11:26.8606732Z  at 
> org.apache.flink.tests.util.AutoClosableProcess$AutoClosableProcessBuilder.runBlockingWithRetry(AutoClosableProcess.java:132)
> 2020-05-05T00:11:26.8607321Z  at 
> org.apache.flink.tests.util.cache.AbstractDownloadCache.getOrDownload(AbstractDownloadCache.java:127)
> 2020-05-05T00:11:26.8607826Z  at 
> org.apache.flink.tests.util.cache.LolCache.getOrDownload(LolCache.java:31)
> 2020-05-05T00:11:26.8608343Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.setupKafkaDist(LocalStandaloneKafkaResource.java:98)
> 2020-05-05T00:11:26.8608892Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.before(LocalStandaloneKafkaResource.java:92)
> 2020-05-05T00:11:26.8609602Z  at 
> org.apache.flink.util.ExternalResource$1.evaluate(ExternalResource.java:46)
> 2020-05-05T00:11:26.8610026Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-05T00:11:26.8610553Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-05T00:11:26.8610958Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-05T00:11:26.8611388Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-05T00:11:26.8612214Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-05T00:11:26.8612706Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-05T00:11:26.8613109Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-05T00:11:26.8613551Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-05T00:11:26.8614019Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-05T00:11:26.8614442Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-05T00:11:26.8614869Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-05T00:11:26.8615251Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-05T00:11:26.8615654Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-05T00:11:26.8616060Z  at 
> 

[jira] [Updated] (FLINK-22333) Elasticsearch7DynamicSinkITCase.testWritingDocuments failed due to deploy task timeout.

2021-04-17 Thread Guowei Ma (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22333?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guowei Ma updated FLINK-22333:
--
Summary: Elasticsearch7DynamicSinkITCase.testWritingDocuments failed due to 
deploy task timeout.  (was: 
Elasticsearch7DynamicSinkITCase.testWritingDocuments fail due to deploy task 
timeout.)

> Elasticsearch7DynamicSinkITCase.testWritingDocuments failed due to deploy 
> task timeout.
> ---
>
> Key: FLINK-22333
> URL: https://issues.apache.org/jira/browse/FLINK-22333
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.13.0
>Reporter: Guowei Ma
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16694=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20=12329
> {code:java}
> 2021-04-16T23:37:23.5719280Z Apr 16 23:37:23 
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
> 2021-04-16T23:37:23.5739250Z Apr 16 23:37:23  at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
> 2021-04-16T23:37:23.5759329Z Apr 16 23:37:23  at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:137)
> 2021-04-16T23:37:23.5779145Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
> 2021-04-16T23:37:23.5799204Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
> 2021-04-16T23:37:23.5819302Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
> 2021-04-16T23:37:23.5839106Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
> 2021-04-16T23:37:23.5859276Z Apr 16 23:37:23  at 
> org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:237)
> 2021-04-16T23:37:23.5868964Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
> 2021-04-16T23:37:23.5869925Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
> 2021-04-16T23:37:23.5919839Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
> 2021-04-16T23:37:23.5959562Z Apr 16 23:37:23  at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
> 2021-04-16T23:37:23.5989732Z Apr 16 23:37:23  at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:1081)
> 2021-04-16T23:37:23.6019422Z Apr 16 23:37:23  at 
> akka.dispatch.OnComplete.internal(Future.scala:264)
> 2021-04-16T23:37:23.6039067Z Apr 16 23:37:23  at 
> akka.dispatch.OnComplete.internal(Future.scala:261)
> 2021-04-16T23:37:23.6060126Z Apr 16 23:37:23  at 
> akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
> 2021-04-16T23:37:23.6089258Z Apr 16 23:37:23  at 
> akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
> 2021-04-16T23:37:23.6119150Z Apr 16 23:37:23  at 
> scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> 2021-04-16T23:37:23.6139149Z Apr 16 23:37:23  at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:73)
> 2021-04-16T23:37:23.6159077Z Apr 16 23:37:23  at 
> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
> 2021-04-16T23:37:23.6189432Z Apr 16 23:37:23  at 
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
> 2021-04-16T23:37:23.6215243Z Apr 16 23:37:23  at 
> akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572)
> 2021-04-16T23:37:23.6219148Z Apr 16 23:37:23  at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
> 2021-04-16T23:37:23.6220221Z Apr 16 23:37:23  at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
> 2021-04-16T23:37:23.6249411Z Apr 16 23:37:23  at 
> scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436)
> 2021-04-16T23:37:23.6259145Z Apr 16 23:37:23  at 
> scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435)
> 2021-04-16T23:37:23.6289272Z Apr 16 23:37:23  at 
> scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> 2021-04-16T23:37:23.6309243Z Apr 16 23:37:23  at 
> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
> 2021-04-16T23:37:23.6359306Z Apr 16 23:37:23  at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
> 2021-04-16T23:37:23.6369399Z Apr 16 23:37:23  at 
> 

[jira] [Created] (FLINK-22333) Elasticsearch7DynamicSinkITCase.testWritingDocuments fail due to deploy task timeout.

2021-04-17 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22333:
-

 Summary: Elasticsearch7DynamicSinkITCase.testWritingDocuments fail 
due to deploy task timeout.
 Key: FLINK-22333
 URL: https://issues.apache.org/jira/browse/FLINK-22333
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.13.0
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16694=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20=12329


{code:java}
2021-04-16T23:37:23.5719280Z Apr 16 23:37:23 
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
2021-04-16T23:37:23.5739250Z Apr 16 23:37:23at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
2021-04-16T23:37:23.5759329Z Apr 16 23:37:23at 
org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:137)
2021-04-16T23:37:23.5779145Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
2021-04-16T23:37:23.5799204Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
2021-04-16T23:37:23.5819302Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
2021-04-16T23:37:23.5839106Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
2021-04-16T23:37:23.5859276Z Apr 16 23:37:23at 
org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:237)
2021-04-16T23:37:23.5868964Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
2021-04-16T23:37:23.5869925Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
2021-04-16T23:37:23.5919839Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
2021-04-16T23:37:23.5959562Z Apr 16 23:37:23at 
java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
2021-04-16T23:37:23.5989732Z Apr 16 23:37:23at 
org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:1081)
2021-04-16T23:37:23.6019422Z Apr 16 23:37:23at 
akka.dispatch.OnComplete.internal(Future.scala:264)
2021-04-16T23:37:23.6039067Z Apr 16 23:37:23at 
akka.dispatch.OnComplete.internal(Future.scala:261)
2021-04-16T23:37:23.6060126Z Apr 16 23:37:23at 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
2021-04-16T23:37:23.6089258Z Apr 16 23:37:23at 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
2021-04-16T23:37:23.6119150Z Apr 16 23:37:23at 
scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
2021-04-16T23:37:23.6139149Z Apr 16 23:37:23at 
org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:73)
2021-04-16T23:37:23.6159077Z Apr 16 23:37:23at 
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
2021-04-16T23:37:23.6189432Z Apr 16 23:37:23at 
scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
2021-04-16T23:37:23.6215243Z Apr 16 23:37:23at 
akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572)
2021-04-16T23:37:23.6219148Z Apr 16 23:37:23at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
2021-04-16T23:37:23.6220221Z Apr 16 23:37:23at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
2021-04-16T23:37:23.6249411Z Apr 16 23:37:23at 
scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436)
2021-04-16T23:37:23.6259145Z Apr 16 23:37:23at 
scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435)
2021-04-16T23:37:23.6289272Z Apr 16 23:37:23at 
scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
2021-04-16T23:37:23.6309243Z Apr 16 23:37:23at 
akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
2021-04-16T23:37:23.6359306Z Apr 16 23:37:23at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
2021-04-16T23:37:23.6369399Z Apr 16 23:37:23at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
2021-04-16T23:37:23.6389444Z Apr 16 23:37:23at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
2021-04-16T23:37:23.6429180Z Apr 16 23:37:23at 
scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
2021-04-16T23:37:23.6449179Z Apr 16 23:37:23at 
akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90)
2021-04-16T23:37:23.6479350Z Apr 16 23:37:23at 

[jira] [Updated] (FLINK-22332) ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse failed due to NPE when calling "notifyDataAvailable"

2021-04-17 Thread Guowei Ma (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22332?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Guowei Ma updated FLINK-22332:
--
Summary: ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse 
failed due to NPE when  calling "notifyDataAvailable"  (was: 
ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse due to NPE 
when  calling "notifyDataAvailable")

> ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse failed due 
> to NPE when  calling "notifyDataAvailable"
> --
>
> Key: FLINK-22332
> URL: https://issues.apache.org/jira/browse/FLINK-22332
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.11.3
>Reporter: Guowei Ma
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=219e462f-e75e-506c-3671-5017d866ccf6=94b2a454-a9e3-5226-421d-758b172639ef=4476



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-22266) Harden JobMasterStopWithSavepointITCase

2021-04-17 Thread Guowei Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22266?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324390#comment-17324390
 ] 

Guowei Ma commented on FLINK-22266:
---

Just for reporting
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16710=logs=8fd9202e-fd17-5b26-353c-ac1ff76c8f28=a0a633b8-47ef-5c5a-2806-3c13b9e48228=4589

> Harden JobMasterStopWithSavepointITCase
> ---
>
> Key: FLINK-22266
> URL: https://issues.apache.org/jira/browse/FLINK-22266
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Runtime / Coordination
>Affects Versions: 1.13.0
>Reporter: Guowei Ma
>Assignee: Dawid Wysakowicz
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.13.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16451=logs=5c8e7682-d68f-54d1-16a2-a09310218a49=f508e270-48d6-5f1e-3138-42a17e0714f0=3884
> {code:java}
> [ERROR] 
> throwingExceptionOnCallbackWithNoRestartsShouldFailTheTerminate(org.apache.flink.runtime.jobmaster.JobMasterStopWithSavepointITCase)
>   Time elapsed: 0.154 s  <<< FAILURE!
> java.lang.AssertionError
>   at org.junit.Assert.fail(Assert.java:86)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at org.junit.Assert.assertTrue(Assert.java:52)
>   at 
> org.apache.flink.runtime.jobmaster.JobMasterStopWithSavepointITCase.throwingExceptionOnCallbackWithoutRestartsHelper(JobMasterStopWithSavepointITCase.java:154)
>   at 
> org.apache.flink.runtime.jobmaster.JobMasterStopWithSavepointITCase.throwingExceptionOnCallbackWithNoRestartsShouldFailTheTerminate(JobMasterStopWithSavepointITCase.java:138)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
>   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-22111) ClientTest.testSimpleRequests fail

2021-04-17 Thread Guowei Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324388#comment-17324388
 ] 

Guowei Ma commented on FLINK-22111:
---

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16710=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20=14747

> ClientTest.testSimpleRequests fail
> --
>
> Key: FLINK-22111
> URL: https://issues.apache.org/jira/browse/FLINK-22111
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Queryable State
>Affects Versions: 1.13.0
>Reporter: Guowei Ma
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16056=logs=3d12d40f-c62d-5ec4-6acc-0efe94cc3e89=5d6e4255-0ea8-5e2a-f52c-c881b7872361=15421
> {code:java}
> 21:47:16,289 [nioEventLoopGroup-4-3] WARN  
> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInitializer [] - 
> Failed to initialize a channel. Closing: [id: 0x40eab0f6, L:/172.29.0.2:43846 
> - R:/172.29.0.2:42436]
> org.apache.flink.shaded.netty4.io.netty.channel.ChannelPipelineException: 
> org.apache.flink.queryablestate.network.ClientTest$1 is not a @Sharable 
> handler, so can't be added or removed multiple times.
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.checkMultiplicity(DefaultChannelPipeline.java:600)
>  ~[flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.addLast(DefaultChannelPipeline.java:202)
>  ~[flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.addLast(DefaultChannelPipeline.java:381)
>  ~[flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.addLast(DefaultChannelPipeline.java:370)
>  ~[flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.queryablestate.network.ClientTest$5.initChannel(ClientTest.java:897)
>  ~[test-classes/:?]
>   at 
> org.apache.flink.queryablestate.network.ClientTest$5.initChannel(ClientTest.java:890)
>  ~[test-classes/:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInitializer.initChannel(ChannelInitializer.java:129)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.ChannelInitializer.handlerAdded(ChannelInitializer.java:112)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.callHandlerAdded(AbstractChannelHandlerContext.java:938)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.callHandlerAdded0(DefaultChannelPipeline.java:609)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.access$100(DefaultChannelPipeline.java:46)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline$PendingHandlerAddedTask.execute(DefaultChannelPipeline.java:1463)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.callHandlerAddedForAllHandlers(DefaultChannelPipeline.java:1115)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.DefaultChannelPipeline.invokeHandlerAddedIfNeeded(DefaultChannelPipeline.java:650)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.register0(AbstractChannel.java:502)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.access$200(AbstractChannel.java:417)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe$1.run(AbstractChannel.java:474)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:500)
>  [flink-shaded-netty-4.1.49.Final-12.0.jar:?]
>   at 
> 

[jira] [Created] (FLINK-22332) ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse due to NPE when calling "notifyDataAvailable"

2021-04-17 Thread Guowei Ma (Jira)
Guowei Ma created FLINK-22332:
-

 Summary: 
ConnectedComponentsWithObjectMapITCase.testJobWithoutObjectReuse due to NPE 
when  calling "notifyDataAvailable"
 Key: FLINK-22332
 URL: https://issues.apache.org/jira/browse/FLINK-22332
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Network
Affects Versions: 1.11.3
Reporter: Guowei Ma


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=219e462f-e75e-506c-3671-5017d866ccf6=94b2a454-a9e3-5226-421d-758b172639ef=4476



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-22307) Increase the default value of data writing cache size (not configurable) for sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao updated FLINK-22307:

Summary: Increase the default value of data writing cache size (not 
configurable) for sort-merge blocking shuffle  (was: Increase the default value 
of data writing cache size for sort-merge blocking shuffle)

> Increase the default value of data writing cache size (not configurable) for 
> sort-merge blocking shuffle
> 
>
> Key: FLINK-22307
> URL: https://issues.apache.org/jira/browse/FLINK-22307
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the data writing cache is 8M, which is not enough if data 
> compression is enabled. By increasing the cache size to 16M, the performance 
> of our benchmark job can be increased by about 20%. (We may make it 
> configurable in the future)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-22307) Increase the default value of data writing cache size for sort-merge blocking shuffle

2021-04-17 Thread Yingjie Cao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yingjie Cao updated FLINK-22307:

Summary: Increase the default value of data writing cache size for 
sort-merge blocking shuffle  (was: Increase the data writing cache size of 
sort-merge blocking shuffle)

> Increase the default value of data writing cache size for sort-merge blocking 
> shuffle
> -
>
> Key: FLINK-22307
> URL: https://issues.apache.org/jira/browse/FLINK-22307
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Network
>Reporter: Yingjie Cao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.13.0
>
>
> Currently, the data writing cache is 8M, which is not enough if data 
> compression is enabled. By increasing the cache size to 16M, the performance 
> of our benchmark job can be increased by about 20%. (We may make it 
> configurable in the future)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-17510) StreamingKafkaITCase. testKafka timeouts on downloading Kafka

2021-04-17 Thread Guowei Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324387#comment-17324387
 ] 

Guowei Ma edited comment on FLINK-17510 at 4/18/21, 2:22 AM:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=2b7514ee-e706-5046-657b-3430666e7bd9=17137

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=68a897ab-3047-5660-245a-cce8f83859f6=d47e27f5-9721-5d5f-1cf3-62adbf3d115d=17334


was (Author: maguowei):
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=2b7514ee-e706-5046-657b-3430666e7bd9=17137


> StreamingKafkaITCase. testKafka timeouts on downloading Kafka
> -
>
> Key: FLINK-17510
> URL: https://issues.apache.org/jira/browse/FLINK-17510
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Connectors / Kafka, Tests
>Affects Versions: 1.11.3, 1.12.1, 1.13.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=585=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-05-05T00:06:49.7268716Z [INFO] 
> ---
> 2020-05-05T00:06:49.7268938Z [INFO]  T E S T S
> 2020-05-05T00:06:49.7269282Z [INFO] 
> ---
> 2020-05-05T00:06:50.5336315Z [INFO] Running 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8603439Z [ERROR] Tests run: 3, Failures: 0, Errors: 2, 
> Skipped: 0, Time elapsed: 276.323 s <<< FAILURE! - in 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8604882Z [ERROR] testKafka[1: 
> kafka-version:0.11.0.2](org.apache.flink.tests.util.kafka.StreamingKafkaITCase)
>   Time elapsed: 120.024 s  <<< ERROR!
> 2020-05-05T00:11:26.8605942Z java.io.IOException: Process ([wget, -q, -P, 
> /tmp/junit2815750531595874769/downloads/1290570732, 
> https://archive.apache.org/dist/kafka/0.11.0.2/kafka_2.11-0.11.0.2.tgz]) 
> exceeded timeout (12) or number of retries (3).
> 2020-05-05T00:11:26.8606732Z  at 
> org.apache.flink.tests.util.AutoClosableProcess$AutoClosableProcessBuilder.runBlockingWithRetry(AutoClosableProcess.java:132)
> 2020-05-05T00:11:26.8607321Z  at 
> org.apache.flink.tests.util.cache.AbstractDownloadCache.getOrDownload(AbstractDownloadCache.java:127)
> 2020-05-05T00:11:26.8607826Z  at 
> org.apache.flink.tests.util.cache.LolCache.getOrDownload(LolCache.java:31)
> 2020-05-05T00:11:26.8608343Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.setupKafkaDist(LocalStandaloneKafkaResource.java:98)
> 2020-05-05T00:11:26.8608892Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.before(LocalStandaloneKafkaResource.java:92)
> 2020-05-05T00:11:26.8609602Z  at 
> org.apache.flink.util.ExternalResource$1.evaluate(ExternalResource.java:46)
> 2020-05-05T00:11:26.8610026Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-05T00:11:26.8610553Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-05T00:11:26.8610958Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-05T00:11:26.8611388Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-05T00:11:26.8612214Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-05T00:11:26.8612706Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-05T00:11:26.8613109Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-05T00:11:26.8613551Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-05T00:11:26.8614019Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-05T00:11:26.8614442Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-05T00:11:26.8614869Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-05T00:11:26.8615251Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-05T00:11:26.8615654Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-05T00:11:26.8616060Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-05T00:11:26.8616465Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-05T00:11:26.8616893Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-05T00:11:26.8617893Z  at 
> 

[jira] [Commented] (FLINK-17510) StreamingKafkaITCase. testKafka timeouts on downloading Kafka

2021-04-17 Thread Guowei Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17510?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324387#comment-17324387
 ] 

Guowei Ma commented on FLINK-17510:
---

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=16711=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=2b7514ee-e706-5046-657b-3430666e7bd9=17137


> StreamingKafkaITCase. testKafka timeouts on downloading Kafka
> -
>
> Key: FLINK-17510
> URL: https://issues.apache.org/jira/browse/FLINK-17510
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Connectors / Kafka, Tests
>Affects Versions: 1.11.3, 1.12.1, 1.13.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=585=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-05-05T00:06:49.7268716Z [INFO] 
> ---
> 2020-05-05T00:06:49.7268938Z [INFO]  T E S T S
> 2020-05-05T00:06:49.7269282Z [INFO] 
> ---
> 2020-05-05T00:06:50.5336315Z [INFO] Running 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8603439Z [ERROR] Tests run: 3, Failures: 0, Errors: 2, 
> Skipped: 0, Time elapsed: 276.323 s <<< FAILURE! - in 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-05-05T00:11:26.8604882Z [ERROR] testKafka[1: 
> kafka-version:0.11.0.2](org.apache.flink.tests.util.kafka.StreamingKafkaITCase)
>   Time elapsed: 120.024 s  <<< ERROR!
> 2020-05-05T00:11:26.8605942Z java.io.IOException: Process ([wget, -q, -P, 
> /tmp/junit2815750531595874769/downloads/1290570732, 
> https://archive.apache.org/dist/kafka/0.11.0.2/kafka_2.11-0.11.0.2.tgz]) 
> exceeded timeout (12) or number of retries (3).
> 2020-05-05T00:11:26.8606732Z  at 
> org.apache.flink.tests.util.AutoClosableProcess$AutoClosableProcessBuilder.runBlockingWithRetry(AutoClosableProcess.java:132)
> 2020-05-05T00:11:26.8607321Z  at 
> org.apache.flink.tests.util.cache.AbstractDownloadCache.getOrDownload(AbstractDownloadCache.java:127)
> 2020-05-05T00:11:26.8607826Z  at 
> org.apache.flink.tests.util.cache.LolCache.getOrDownload(LolCache.java:31)
> 2020-05-05T00:11:26.8608343Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.setupKafkaDist(LocalStandaloneKafkaResource.java:98)
> 2020-05-05T00:11:26.8608892Z  at 
> org.apache.flink.tests.util.kafka.LocalStandaloneKafkaResource.before(LocalStandaloneKafkaResource.java:92)
> 2020-05-05T00:11:26.8609602Z  at 
> org.apache.flink.util.ExternalResource$1.evaluate(ExternalResource.java:46)
> 2020-05-05T00:11:26.8610026Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-05T00:11:26.8610553Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-05T00:11:26.8610958Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-05T00:11:26.8611388Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-05T00:11:26.8612214Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-05T00:11:26.8612706Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-05T00:11:26.8613109Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-05T00:11:26.8613551Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-05T00:11:26.8614019Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-05T00:11:26.8614442Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-05T00:11:26.8614869Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-05T00:11:26.8615251Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-05T00:11:26.8615654Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-05T00:11:26.8616060Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-05T00:11:26.8616465Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-05T00:11:26.8616893Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-05T00:11:26.8617893Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-05T00:11:26.8618490Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-05T00:11:26.8619056Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-05T00:11:26.8619589Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-05T00:11:26.8620073Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 

[GitHub] [flink] wsry commented on a change in pull request #15652: [FLINK-22305][network] Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread GitBox


wsry commented on a change in pull request #15652:
URL: https://github.com/apache/flink/pull/15652#discussion_r615329427



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/SortMergeResultPartition.java
##
@@ -160,18 +160,12 @@ public void setup() throws IOException {
 }
 
 int numRequiredBuffer = bufferPool.getNumberOfRequiredMemorySegments();
-String errorMessage =
-String.format(
-"Too few sort buffers, please increase %s to a larger 
value (more than %d).",
-
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS,
-2 * expectedWriteBuffers);
-if (numRequiredBuffer < 2 * expectedWriteBuffers) {
-LOG.warn(errorMessage);
-}
-
 int numWriteBuffers = Math.min(numRequiredBuffer / 2, 
expectedWriteBuffers);
 if (numWriteBuffers < 1) {
-throw new IOException(errorMessage);
+throw new IOException(
+String.format(
+"Too few sort buffers, please increase %s.",
+
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS));

Review comment:
   The error only occurs when only 1 buffer is configured which is rare and 
increasing it to any value would be OK. Besides, we already give a suggested 
value in the config option description which is good enough and should never 
trigger any warnings or exceptions.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] guoweiM commented on a change in pull request #15652: [FLINK-22305][network] Improve log messages of sort-merge blocking shuffle

2021-04-17 Thread GitBox


guoweiM commented on a change in pull request #15652:
URL: https://github.com/apache/flink/pull/15652#discussion_r615328699



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/SortMergeResultPartition.java
##
@@ -160,18 +160,12 @@ public void setup() throws IOException {
 }
 
 int numRequiredBuffer = bufferPool.getNumberOfRequiredMemorySegments();
-String errorMessage =
-String.format(
-"Too few sort buffers, please increase %s to a larger 
value (more than %d).",
-
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS,
-2 * expectedWriteBuffers);
-if (numRequiredBuffer < 2 * expectedWriteBuffers) {
-LOG.warn(errorMessage);
-}
-
 int numWriteBuffers = Math.min(numRequiredBuffer / 2, 
expectedWriteBuffers);
 if (numWriteBuffers < 1) {
-throw new IOException(errorMessage);
+throw new IOException(
+String.format(
+"Too few sort buffers, please increase %s.",
+
NettyShuffleEnvironmentOptions.NETWORK_SORT_SHUFFLE_MIN_BUFFERS));

Review comment:
   I think this it might not be clearer than the before. The old message 
tells user increase to what size of the network buffer but the new message does 
not tell the size of network buffer that user should increase to.




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] link3280 commented on a change in pull request #13377: [FLINK-18592] bugfix for StreamingFileSink

2021-04-17 Thread GitBox


link3280 commented on a change in pull request #13377:
URL: https://github.com/apache/flink/pull/13377#discussion_r615326763



##
File path: 
flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopRecoverableFsDataOutputStream.java
##
@@ -360,4 +366,121 @@ private static boolean waitUntilLeaseIsRevoked(final 
FileSystem fs, final Path p
 }
 return isClosed;
 }
+
+
+   /*
+* Run the dfs recover lease. recoverLease is asynchronous. It returns: 
-false when it starts the lease recovery (i.e. lease recovery not *yet* done) - 
true when the lease recovery has
+* succeeded or the file is closed.
+*
+* But, we have to be careful. Each time we call recoverLease, it 
starts the recover lease process over from the beginning. We could put 
ourselves in a situation
+* where we are doing nothing but starting a recovery, interrupting it 
to start again, and so on.
+*
+* The namenode will try to recover the lease on the file's primary 
node. If all is well, it should return near immediately.
+* But, as is common, it is the very primary node that has crashed and 
so the namenode will be stuck waiting on a socket timeout before it will ask 
another datanode to start the recovery.
+* It does not help if we call recoverLease in the meantime and in 
particular, subsequent to the socket timeout, a recoverLease invocation will 
cause us to start over from square one
+* (possibly waiting on socket timeout against primary node).
+* So, in the below, we do the following:
+* 1. Call recoverLease.
+* 2. If it returns true, break.
+* 3. If it returns false, wait a few seconds and then call it again.
+* 4. If it returns true, break.
+* 5. If it returns false, wait for what we think the datanode socket 
timeout is (configurable) and then try again.
+* 6. If it returns true, break.
+* 7. If it returns false, repeat starting at step 5. above. If 
HDFS-4525 is available, call it every second and we might be able to exit early.
+*/
+   private static boolean recoverLease(Path path, DistributedFileSystem 
dfs) throws IOException {

Review comment:
   Looks like this new method is unused. Shouldn't we replace 
#waitUntilLeaseIsRevoked with it? 

##
File path: 
flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopRecoverableFsDataOutputStream.java
##
@@ -357,6 +364,126 @@ private static boolean waitUntilLeaseIsRevoked(final 
FileSystem fs, final Path p
}
isClosed = dfs.isFileClosed(path);
}
+   // [FLINK-18592] recover lease after the lease timeout passed 
but file was still not closed
+   if(!isClosed && !deadline.hasTimeLeft()){
+   recoverLease(path, dfs);
+   }
return isClosed;
}
+
+
+   /*
+* Run the dfs recover lease. recoverLease is asynchronous. It returns: 
-false when it starts the lease recovery (i.e. lease recovery not *yet* done) - 
true when the lease recovery has
+* succeeded or the file is closed.
+*
+* But, we have to be careful. Each time we call recoverLease, it 
starts the recover lease process over from the beginning. We could put 
ourselves in a situation
+* where we are doing nothing but starting a recovery, interrupting it 
to start again, and so on.
+*
+* The namenode will try to recover the lease on the file's primary 
node. If all is well, it should return near immediately.
+* But, as is common, it is the very primary node that has crashed and 
so the namenode will be stuck waiting on a socket timeout before it will ask 
another datanode to start the recovery.
+* It does not help if we call recoverLease in the meantime and in 
particular, subsequent to the socket timeout, a recoverLease invocation will 
cause us to start over from square one
+* (possibly waiting on socket timeout against primary node).
+* So, in the below, we do the following:
+* 1. Call recoverLease.
+* 2. If it returns true, break.
+* 3. If it returns false, wait a few seconds and then call it again.
+* 4. If it returns true, break.
+* 5. If it returns false, wait for what we think the datanode socket 
timeout is (configurable) and then try again.
+* 6. If it returns true, break.
+* 7. If it returns false, repeat starting at step 5. above. If 
HDFS-4525 is available, call it every second and we might be able to exit early.
+*/
+   private static boolean recoverLease(Path path, DistributedFileSystem 
dfs) throws IOException {
+   LOG.info("Recover lease on dfs file " + path);
+   long startWaiting = System.currentTimeMillis();
+   

[GitHub] [flink] flinkbot edited a comment on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821836292


   
   ## CI report:
   
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821836292


   
   ## CI report:
   
   *  Unknown: [CANCELED](TBD) 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] brandonJY commented on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


brandonJY commented on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821911681


   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] brandonJY commented on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


brandonJY commented on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821911588


   @flinkbot run travis


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15574: [hotfix][docs] Removed reference to fold in window function

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15574:
URL: https://github.com/apache/flink/pull/15574#issuecomment-817750003


   
   ## CI report:
   
   * fba7c0da981277648ff3629c1f4de48d1ccdae45 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16713)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15574: [hotfix][docs] Removed reference to fold in window function

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15574:
URL: https://github.com/apache/flink/pull/15574#issuecomment-817750003


   
   ## CI report:
   
   * 6e8a34bd5b59ebf1afa8149edc9f3ca232c35bf2 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16388)
 
   * fba7c0da981277648ff3629c1f4de48d1ccdae45 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16713)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15574: [hotfix][docs] Removed reference to fold in window function

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15574:
URL: https://github.com/apache/flink/pull/15574#issuecomment-817750003


   
   ## CI report:
   
   * 6e8a34bd5b59ebf1afa8149edc9f3ca232c35bf2 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16388)
 
   * fba7c0da981277648ff3629c1f4de48d1ccdae45 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15121: The method $(String) is undefined for the type TableExample

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15121:
URL: https://github.com/apache/flink/pull/15121#issuecomment-793480240


   
   ## CI report:
   
   * cf26ce895a7956d258acc073817f578558e78227 UNKNOWN
   * 709c4110370b845f42d916a06e56c6026cf2fac8 UNKNOWN
   * 4ea99332e1997eebca1f3f0a9d9229b8265fe32c UNKNOWN
   * d39b3fb096a3c883d36d3dcc85ea93efda8edd3d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16708)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-7883) Make savepoints atomic with respect to state and side effects

2021-04-17 Thread Olivier NOUGUIER (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-7883?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324335#comment-17324335
 ] 

Olivier NOUGUIER commented on FLINK-7883:
-

Any update ?

 

> Make savepoints atomic with respect to state and side effects
> -
>
> Key: FLINK-7883
> URL: https://issues.apache.org/jira/browse/FLINK-7883
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataStream, Connectors / Kafka, Runtime / State 
> Backends
>Affects Versions: 1.3.2, 1.4.0
>Reporter: Antoine Philippot
>Priority: Major
>
> For a cancel with savepoint command, the JobManager trigger the cancel call 
> once the savepoint is finished, but during the savepoint execution, kafka 
> source continue to poll new messages which will not be part of the savepoint 
> and will be replayed on the next application start.
> A solution could be to stop fetching the source stream task before triggering 
> the savepoint.
> I suggest to add an interface {{StoppableFetchingSourceFunction}} with a 
> method {{stopFetching}} that existant SourceFunction implementations could 
> implement.
> We can add a {{stopFetchingSource}} property in 
>  {{CheckpointOptions}} class to pass the desired behaviour from 
> {{JobManager.handleMessage(CancelJobWithSavepoint)}} to 
> {{SourceStreamTask.triggerCheckpoint}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-22331) CLI Frontend (RestClient) doesn't work on Apple M1

2021-04-17 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-22331:
--

 Summary: CLI Frontend (RestClient) doesn't work on Apple M1
 Key: FLINK-22331
 URL: https://issues.apache.org/jira/browse/FLINK-22331
 Project: Flink
  Issue Type: Bug
  Components: Client / Job Submission
Affects Versions: 1.12.2, 1.13.0
Reporter: Robert Metzger
 Attachments: flink-muthmann-client-KlemensMac.local (1).log_native, 
flink-muthmann-client-KlemensMac.local.log_rosetta

This issue was first reported by a user: 
https://lists.apache.org/thread.html/r50bda40a69688de52c9d6e3489ac2641491387c20fdc1cecedceee76%40%3Cuser.flink.apache.org%3E

See attached logs.

Exception without rosetta:
{code}
org.apache.flink.client.program.ProgramInvocationException: The main method 
caused an error: Failed to execute job 'Streaming WordCount'.
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:366)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:219)
at 
org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:114)
at 
org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:812)
at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:246)
at 
org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1054)
at 
org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
Caused by: org.apache.flink.util.FlinkException: Failed to execute job 
'Streaming WordCount'.
at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1918)
at 
org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:135)
at 
org.apache.flink.client.program.StreamContextEnvironment.execute(StreamContextEnvironment.java:76)
at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1782)
at 
org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCount.java:97)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:349)
... 8 more
Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed to 
submit JobGraph.
at 
org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:400)
at 
java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
at 
java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
at 
org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$9(FutureUtils.java:390)
at 
java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
at 
java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
at 
java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
at 
java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088)
at 
org.apache.flink.runtime.rest.RestClient$ClientHandler.exceptionCaught(RestClient.java:613)
at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:302)
at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.invokeExceptionCaught(AbstractChannelHandlerContext.java:281)
at 
org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannelHandlerContext.fireExceptionCaught(AbstractChannelHandlerContext.java:273)
at 
org.apache.flink.shaded.netty4.io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireExceptionCaught(CombinedChannelDuplexHandler.java:424)
at 

[GitHub] [flink] flinkbot edited a comment on pull request #15534: [FLINK-22104][sql-client] Fix unstable SqlClientTest.testExecuteSqlFile

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15534:
URL: https://github.com/apache/flink/pull/15534#issuecomment-816392253


   
   ## CI report:
   
   * f0f37cfe199dc670e8908b05a7df7ace8c2ecfd4 UNKNOWN
   * 556c5fe14574005af9f6937c6ab4e1a5dc888d29 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16706)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] fsk119 commented on pull request #15585: [FLINK-22065] [sql-client]Beautify the parse error exception

2021-04-17 Thread GitBox


fsk119 commented on pull request #15585:
URL: https://github.com/apache/flink/pull/15585#issuecomment-821850442


   There are check style  errors. Use `mvn 
com.diffplug.spotless:spotless-maven-plugin:apply` to fix these.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] fsk119 commented on a change in pull request #15585: [FLINK-22065] [sql-client]Beautify the parse error exception

2021-04-17 Thread GitBox


fsk119 commented on a change in pull request #15585:
URL: https://github.com/apache/flink/pull/15585#discussion_r615272815



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/exception/SqlParseException.java
##
@@ -0,0 +1,32 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.client.exception;
+
+/** Exception thrown during the parse of SQL statements. */
+public class SqlParseException extends SqlClientException {
+private static final long serialVersionUID = 1L;
+
+public SqlParseException(String message) {
+super(message);
+}

Review comment:
   This is method is not used. Remove this.

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/CliClient.java
##
@@ -320,8 +324,8 @@ private boolean executeStatement(String statement, 
ExecutionMode executionMode)
 try {
 final Optional operation = parseCommand(statement);
 operation.ifPresent(op -> callOperation(op, executionMode));
-} catch (SqlExecutionException e) {
-printExecutionException(e);
+} catch (SqlClientException e) {
+printSqlClientException(e);
 return false;

Review comment:
   What about 
   ```
   catch (SqlParseException e) {
  printSqlParseException(e);
   }
   catch (SqlExecutionException e) {
  printSqlExecutionException(e);
   }
   ```
   We can remove the `instance of`.
   

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/gateway/local/LocalExecutor.java
##
@@ -175,10 +176,10 @@ public Operation parseStatement(String sessionId, String 
statement)
 try {
 operations = context.wrapClassLoader(() -> 
parser.parse(statement));
 } catch (Exception e) {
-throw new SqlExecutionException("Failed to parse statement: " + 
statement, e);
+throw new SqlParseException("Failed to parse statement: " + 
statement, e);
 }
 if (operations.isEmpty()) {
-throw new SqlExecutionException("Failed to parse statement: " + 
statement);
+throw new SqlExecutionException("invalid statement: " + statement);

Review comment:
   Use `SqlParseException`? I think the origin exception message is good 
enough. 

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/gateway/Executor.java
##
@@ -69,74 +70,71 @@
  * Both this method and {@link #getSessionConfigMap(String)} return the 
same configuration
  * set, but different return type.
  */
-ReadableConfig getSessionConfig(String sessionId) throws 
SqlExecutionException;
+ReadableConfig getSessionConfig(String sessionId) throws 
SqlClientException;
 
 /**
  * Reset all the properties for the given session identifier.
  *
  * @param sessionId to identifier the session
- * @throws SqlExecutionException if any error happen.
+ * @throws SqlClientException if any error happen.
  */
-void resetSessionProperties(String sessionId) throws SqlExecutionException;
+void resetSessionProperties(String sessionId) throws SqlClientException;
 
 /**
  * Reset given key's the session property for default value, if key is not 
defined in config
  * file, then remove it.
  *
  * @param sessionId to identifier the session
  * @param key of need to reset the session property
- * @throws SqlExecutionException if any error happen.
+ * @throws SqlClientException if any error happen.
  */
-void resetSessionProperty(String sessionId, String key) throws 
SqlExecutionException;
+void resetSessionProperty(String sessionId, String key) throws 
SqlClientException;
 
 /**
  * Set given key's session property to the specific value.
  *
  * @param key of the session property
  * @param value of the session property
- * @throws SqlExecutionException if any error happen.
+ * @throws SqlClientException if any error happen.
  */
-void setSessionProperty(String sessionId, String key, String value)
- 

[GitHub] [flink] flinkbot edited a comment on pull request #15483: [FLINK-22092][hive] Ignore static conf file URLs in HiveConf

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15483:
URL: https://github.com/apache/flink/pull/15483#issuecomment-812520117


   
   ## CI report:
   
   * b91b371f5b6a9b522c2478ec56dfe48231c9c976 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16705)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821836292


   
   ## CI report:
   
   * ab75719fea8ed2bc179cfe18557211fd27510f3f Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16709)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15121: The method $(String) is undefined for the type TableExample

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15121:
URL: https://github.com/apache/flink/pull/15121#issuecomment-793480240


   
   ## CI report:
   
   * cf26ce895a7956d258acc073817f578558e78227 UNKNOWN
   * 709c4110370b845f42d916a06e56c6026cf2fac8 UNKNOWN
   * 4ea99332e1997eebca1f3f0a9d9229b8265fe32c UNKNOWN
   * 902a0f54b487ee39a8adb0e04c826bcc4a9dcab7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16641)
 
   * d39b3fb096a3c883d36d3dcc85ea93efda8edd3d Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16708)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] AHeise commented on a change in pull request #15631: [FLINK-20103][tests] Add UnalignedCheckpointsStressITCase

2021-04-17 Thread GitBox


AHeise commented on a change in pull request #15631:
URL: https://github.com/apache/flink/pull/15631#discussion_r615265639



##
File path: 
flink-tests/src/test/java/org/apache/flink/test/checkpointing/UnalignedCheckpointStressITCase.java
##
@@ -0,0 +1,553 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.test.checkpointing;
+
+import org.apache.flink.api.common.functions.AbstractRichFunction;
+import org.apache.flink.api.common.functions.MapFunction;
+import org.apache.flink.api.common.restartstrategy.RestartStrategies;
+import org.apache.flink.api.common.state.CheckpointListener;
+import org.apache.flink.api.common.state.ListState;
+import org.apache.flink.api.common.state.ListStateDescriptor;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.RestOptions;
+import org.apache.flink.runtime.jobgraph.JobGraph;
+import org.apache.flink.runtime.jobgraph.SavepointRestoreSettings;
+import org.apache.flink.runtime.operators.testutils.ExpectedTestException;
+import org.apache.flink.runtime.state.FunctionInitializationContext;
+import org.apache.flink.runtime.state.FunctionSnapshotContext;
+import org.apache.flink.runtime.testutils.MiniClusterResourceConfiguration;
+import org.apache.flink.streaming.api.checkpoint.CheckpointedFunction;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.streaming.api.datastream.DataStreamSource;
+import org.apache.flink.streaming.api.datastream.DataStreamUtils;
+import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
+import 
org.apache.flink.streaming.api.environment.CheckpointConfig.ExternalizedCheckpointCleanup;
+import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
+import org.apache.flink.streaming.api.functions.co.KeyedCoProcessFunction;
+import org.apache.flink.streaming.api.functions.source.ParallelSourceFunction;
+import 
org.apache.flink.streaming.api.functions.windowing.ProcessWindowFunction;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows;
+import org.apache.flink.streaming.api.windowing.time.Time;
+import org.apache.flink.streaming.api.windowing.windows.TimeWindow;
+import org.apache.flink.test.util.MiniClusterWithClientResource;
+import org.apache.flink.util.Collector;
+import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.FileUtils;
+import org.apache.flink.util.TestLogger;
+
+import org.junit.After;
+import org.junit.Before;
+import org.junit.ClassRule;
+import org.junit.Test;
+import org.junit.rules.TemporaryFolder;
+
+import javax.annotation.Nullable;
+
+import java.io.File;
+import java.io.IOException;
+import java.nio.file.Files;
+import java.nio.file.Path;
+import java.nio.file.Paths;
+import java.util.Comparator;
+import java.util.List;
+import java.util.Optional;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.regex.Matcher;
+import java.util.regex.Pattern;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import static 
org.apache.flink.configuration.CheckpointingOptions.CHECKPOINTS_DIRECTORY;
+import static 
org.apache.flink.configuration.CheckpointingOptions.MAX_RETAINED_CHECKPOINTS;
+import static 
org.apache.flink.shaded.guava18.com.google.common.collect.Iterables.getOnlyElement;
+import static org.apache.flink.test.util.TestUtils.submitJobAndWaitForResult;
+import static org.apache.flink.util.Preconditions.checkState;
+
+/**
+ * A stress test that runs for a pre-defined amount of time, verifying data 
correctness and every
+ * couple of checkpoints is triggering fail over to stress test unaligned 
checkpoints.
+ */
+@SuppressWarnings("serial")
+public class UnalignedCheckpointStressITCase extends TestLogger {
+
+protected static final int CHECKPOINT_INTERVAL = 20;
+protected static final int MINIMUM_COMPLETED_CHECKPOINTS_BETWEEN_FAILURES 
= 2;
+protected static final int MAXIMUM_COMPLETED_CHECKPOINTS_BETWEEN_FAILURES 
= 10;
+protected static final long TEST_DURATION = 
Time.minutes(1).toMilliseconds();
+protected static final int 

[GitHub] [flink] AHeise commented on a change in pull request #15631: [FLINK-20103][tests] Add UnalignedCheckpointsStressITCase

2021-04-17 Thread GitBox


AHeise commented on a change in pull request #15631:
URL: https://github.com/apache/flink/pull/15631#discussion_r615264959



##
File path: 
flink-tests/src/test/java/org/apache/flink/test/checkpointing/UnalignedCheckpointStressITCase.java
##
@@ -0,0 +1,553 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.test.checkpointing;
+
+import org.apache.flink.api.common.functions.AbstractRichFunction;
+import org.apache.flink.api.common.functions.MapFunction;
+import org.apache.flink.api.common.restartstrategy.RestartStrategies;
+import org.apache.flink.api.common.state.CheckpointListener;
+import org.apache.flink.api.common.state.ListState;
+import org.apache.flink.api.common.state.ListStateDescriptor;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.RestOptions;
+import org.apache.flink.runtime.jobgraph.JobGraph;
+import org.apache.flink.runtime.jobgraph.SavepointRestoreSettings;
+import org.apache.flink.runtime.operators.testutils.ExpectedTestException;
+import org.apache.flink.runtime.state.FunctionInitializationContext;
+import org.apache.flink.runtime.state.FunctionSnapshotContext;
+import org.apache.flink.runtime.testutils.MiniClusterResourceConfiguration;
+import org.apache.flink.streaming.api.checkpoint.CheckpointedFunction;
+import org.apache.flink.streaming.api.datastream.DataStream;
+import org.apache.flink.streaming.api.datastream.DataStreamSource;
+import org.apache.flink.streaming.api.datastream.DataStreamUtils;
+import org.apache.flink.streaming.api.datastream.SingleOutputStreamOperator;
+import 
org.apache.flink.streaming.api.environment.CheckpointConfig.ExternalizedCheckpointCleanup;
+import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
+import org.apache.flink.streaming.api.functions.co.KeyedCoProcessFunction;
+import org.apache.flink.streaming.api.functions.source.ParallelSourceFunction;
+import 
org.apache.flink.streaming.api.functions.windowing.ProcessWindowFunction;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows;
+import org.apache.flink.streaming.api.windowing.time.Time;
+import org.apache.flink.streaming.api.windowing.windows.TimeWindow;
+import org.apache.flink.test.util.MiniClusterWithClientResource;
+import org.apache.flink.util.Collector;
+import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.FileUtils;
+import org.apache.flink.util.TestLogger;
+
+import org.junit.After;
+import org.junit.Before;
+import org.junit.ClassRule;
+import org.junit.Test;
+import org.junit.rules.TemporaryFolder;
+
+import javax.annotation.Nullable;
+
+import java.io.File;
+import java.io.IOException;
+import java.nio.file.Files;
+import java.nio.file.Path;
+import java.nio.file.Paths;
+import java.util.Comparator;
+import java.util.List;
+import java.util.Optional;
+import java.util.concurrent.ThreadLocalRandom;
+import java.util.regex.Matcher;
+import java.util.regex.Pattern;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import static 
org.apache.flink.configuration.CheckpointingOptions.CHECKPOINTS_DIRECTORY;
+import static 
org.apache.flink.configuration.CheckpointingOptions.MAX_RETAINED_CHECKPOINTS;
+import static 
org.apache.flink.shaded.guava18.com.google.common.collect.Iterables.getOnlyElement;
+import static org.apache.flink.test.util.TestUtils.submitJobAndWaitForResult;
+import static org.apache.flink.util.Preconditions.checkState;
+
+/**
+ * A stress test that runs for a pre-defined amount of time, verifying data 
correctness and every
+ * couple of checkpoints is triggering fail over to stress test unaligned 
checkpoints.
+ */
+@SuppressWarnings("serial")
+public class UnalignedCheckpointStressITCase extends TestLogger {
+
+protected static final int CHECKPOINT_INTERVAL = 20;
+protected static final int MINIMUM_COMPLETED_CHECKPOINTS_BETWEEN_FAILURES 
= 2;
+protected static final int MAXIMUM_COMPLETED_CHECKPOINTS_BETWEEN_FAILURES 
= 10;
+protected static final long TEST_DURATION = 
Time.minutes(1).toMilliseconds();
+protected static final int 

[GitHub] [flink] flinkbot commented on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


flinkbot commented on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821836292


   
   ## CI report:
   
   * ab75719fea8ed2bc179cfe18557211fd27510f3f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15653: [FLINK-22329][hive] Inject current ugi credentials into jobconf when getting file split in hive connector

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15653:
URL: https://github.com/apache/flink/pull/15653#issuecomment-821801891


   
   ## CI report:
   
   * 41dbfa9649149d9616337cccf4fc8f6c6f59a9c5 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16704)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-22314) AggRecordsCombiner should combine buffered records first instead of accumulate on state directly

2021-04-17 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-22314?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-22314.
---
Resolution: Invalid

> AggRecordsCombiner should combine buffered records first instead of 
> accumulate on state directly
> 
>
> Key: FLINK-22314
> URL: https://issues.apache.org/jira/browse/FLINK-22314
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Reporter: Jark Wu
>Assignee: Jark Wu
>Priority: Major
>
> In Window TVF Aggregation, currently, the {{AggRecordsCombiner}} accumulates 
> buffered records on state directly. This is not good for performance. We can 
> accumulate records in memory first, and then merge the accumulator into 
> state, if the aggs support {{merge()}} method. This can reduce lots of state 
> accessing when having {{COUNT DISTINCT}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong commented on a change in pull request #13011: [FLINK-16384][table sql/client] Support SHOW CREATE TABLE statement

2021-04-17 Thread GitBox


wuchong commented on a change in pull request #13011:
URL: https://github.com/apache/flink/pull/13011#discussion_r465552827



##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/api/constraints/UniqueConstraint.java
##
@@ -75,21 +76,26 @@ public ConstraintType getType() {
 */
@Override
public final String asSummaryString() {
-   final String typeString;
-   switch (getType()) {
-   case PRIMARY_KEY:
-   typeString = "PRIMARY KEY";
-   break;
-   case UNIQUE_KEY:
-   typeString = "UNIQUE";
-   break;
-   default:
-   throw new IllegalStateException("Unknown key 
type: " + getType());
-   }
-
+   final String typeString = getTypeString();
return String.format("CONSTRAINT %s %s (%s)", getName(), 
typeString, String.join(", ", columns));
}
 
+   /**
+* Returns constraint's canonical summary. All constraints summary will 
be formatted as
+* 
+* CONSTRAINT [constraint-name] [constraint-type] 
([constraint-definition]) NOT ENFORCED
+*
+* E.g CONSTRAINT pk PRIMARY KEY (`f0`, `f1`) NOT ENFORCED
+* 
+*/
+   public final String asCanonicalString() {
+   final String typeString = getTypeString();
+   return String.format("CONSTRAINT %s %s (%s) NOT ENFORCED",
+   getName(),
+   typeString,
+   String.join(", ", columns.stream().map(col -> 
String.format("`%s`", col)).collect(Collectors.toList(;

Review comment:
   Use `EncodingUtils#escapeIdentifier` to escape identifiers. 

##
File path: docs/content/docs/dev/table/sql/show.md
##
@@ -427,6 +458,17 @@ SHOW TABLES
 
 Show all tables in the current catalog and the current database.
 
+
+## SHOW CREATE TABLE
+
+```sql
+SHOW CREATE TABLE
+```
+
+Show create table statement for specified table.
+
+Attention Currently `SHOW CREATE 
TABLE` only supports table that is created by Flink SQL. 

Review comment:
   minor:
   
   ```suggestion
   Attention Currently `SHOW CREATE 
TABLE` only supports table that is created by Flink SQL DDL. 
   ```

##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/api/constraints/UniqueConstraint.java
##
@@ -75,21 +76,26 @@ public ConstraintType getType() {
 */
@Override
public final String asSummaryString() {
-   final String typeString;
-   switch (getType()) {
-   case PRIMARY_KEY:
-   typeString = "PRIMARY KEY";
-   break;
-   case UNIQUE_KEY:
-   typeString = "UNIQUE";
-   break;
-   default:
-   throw new IllegalStateException("Unknown key 
type: " + getType());
-   }
-
+   final String typeString = getTypeString();
return String.format("CONSTRAINT %s %s (%s)", getName(), 
typeString, String.join(", ", columns));

Review comment:
   I think we should also add `NOT ENFORCED` for the summary string. 
Currentlly, it is not correct. 
   Besides, we should add `NOT ENFORCED` according the underlying `enforced` 
flag, even though it is always flase for now. 

##
File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/internal/TableEnvironmentImpl.java
##
@@ -1072,6 +1086,75 @@ private TableResult buildShowResult(String columnName, 
String[] objects) {
Arrays.stream(objects).map((c) -> new 
String[]{c}).toArray(String[][]::new));
}
 
+   private TableResult buildShowCreateTableResult(CatalogBaseTable table, 
ObjectIdentifier sqlIdentifier) {

Review comment:
   I think this must be a `CatalogTable`, because this is `SHOW CREATE 
TABLE`.  We should throw exception before this if it is a view. 

##
File path: 
flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/api/internal/TableEnvironmentImpl.java
##
@@ -1015,6 +1018,17 @@ private TableResult executeOperation(Operation 
operation) {
return buildShowResult("database name", 
listDatabases());
} else if (operation instanceof ShowCurrentDatabaseOperation) {
return buildShowResult("current database name", new 
String[]{catalogManager.getCurrentDatabase()});
+   } else if (operation instanceof ShowCreateTableOperation) {
+   ShowCreateTableOperation showCreateTableOperation = 
(ShowCreateTableOperation) operation;
+   Optional 

[GitHub] [flink] flinkbot edited a comment on pull request #15534: [FLINK-22104][sql-client] Fix unstable SqlClientTest.testExecuteSqlFile

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15534:
URL: https://github.com/apache/flink/pull/15534#issuecomment-816392253


   
   ## CI report:
   
   * f0f37cfe199dc670e8908b05a7df7ace8c2ecfd4 UNKNOWN
   * 8a566552f9d186ed1736686f60f13f943dc70838 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16680)
 
   * 556c5fe14574005af9f6937c6ab4e1a5dc888d29 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16706)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15121: The method $(String) is undefined for the type TableExample

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15121:
URL: https://github.com/apache/flink/pull/15121#issuecomment-793480240


   
   ## CI report:
   
   * cf26ce895a7956d258acc073817f578558e78227 UNKNOWN
   * 709c4110370b845f42d916a06e56c6026cf2fac8 UNKNOWN
   * 4ea99332e1997eebca1f3f0a9d9229b8265fe32c UNKNOWN
   * 902a0f54b487ee39a8adb0e04c826bcc4a9dcab7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16641)
 
   * d39b3fb096a3c883d36d3dcc85ea93efda8edd3d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


flinkbot commented on pull request #15654:
URL: https://github.com/apache/flink/pull/15654#issuecomment-821829689


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit ab75719fea8ed2bc179cfe18557211fd27510f3f (Sat Apr 17 
14:17:07 UTC 2021)
   
   **Warnings:**
* **1 pom.xml files were touched**: Check for build and licensing issues.
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-22330) Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread Brandon Jiang (Jira)
Brandon Jiang created FLINK-22330:
-

 Summary: Bumps checkstyle, junit, log4j version for security
 Key: FLINK-22330
 URL: https://issues.apache.org/jira/browse/FLINK-22330
 Project: Flink
  Issue Type: Improvement
Affects Versions: 1.12.2
Reporter: Brandon Jiang






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] brandonJY opened a new pull request #15654: [FLINK-22141] Bumps checkstyle, junit, log4j version for security

2021-04-17 Thread GitBox


brandonJY opened a new pull request #15654:
URL: https://github.com/apache/flink/pull/15654


   
   
   
   
   
   
   ## What is the purpose of the change
   
   
   - Bump junit from 4.12 to 4.13.1 due to GHSA-269g-pwp5-87pp 
https://github.com/advisories/GHSA-269g-pwp5-87pp
   - Bump log4j.version from 2.12.1 to 2.13.3 due to CVE-2020-9488 
https://github.com/advisories/GHSA-vwqq-5vrc-xw9h
   - Bump Bump checkstyle from 8.14 to 8.29 due to GHSA-763g-fqq7-48wg 
https://github.com/advisories/GHSA-763g-fqq7-48wg
   
   ## Brief change log
   
   - Bump junit from 4.12 to 4.13.1 due to GHSA-269g-pwp5-87pp 
https://github.com/advisories/GHSA-269g-pwp5-87pp
   - Bump log4j.version from 2.12.1 to 2.13.3 due to CVE-2020-9488 
https://github.com/advisories/GHSA-vwqq-5vrc-xw9h
   - Bump Bump checkstyle from 8.14 to 8.29 due to GHSA-763g-fqq7-48wg 
https://github.com/advisories/GHSA-763g-fqq7-48wg
   
   
   ## Verifying this change
   
   
   
   This change is already covered by existing tests,
   
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes )
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: ( no)
 - The serializers: (yes)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes)
 - The S3 file system connector: (don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #15534: [FLINK-22104][sql-client] Fix unstable SqlClientTest.testExecuteSqlFile

2021-04-17 Thread GitBox


flinkbot edited a comment on pull request #15534:
URL: https://github.com/apache/flink/pull/15534#issuecomment-816392253


   
   ## CI report:
   
   * f0f37cfe199dc670e8908b05a7df7ace8c2ecfd4 UNKNOWN
   * 8a566552f9d186ed1736686f60f13f943dc70838 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=16680)
 
   * 556c5fe14574005af9f6937c6ab4e1a5dc888d29 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong commented on a change in pull request #15559: [FLINK-21627][table-planner-blink] The digest of TableScan & Sink should consider table hints

2021-04-17 Thread GitBox


wuchong commented on a change in pull request #15559:
URL: https://github.com/apache/flink/pull/15559#discussion_r615256245



##
File path: 
flink-table/flink-table-planner-blink/src/main/java/org/apache/calcite/rel/logical/LogicalTableScan.java
##
@@ -0,0 +1,135 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.calcite.rel.logical;
+
+import com.google.common.collect.ImmutableList;
+import org.apache.calcite.plan.Convention;
+import org.apache.calcite.plan.RelOptCluster;
+import org.apache.calcite.plan.RelOptTable;
+import org.apache.calcite.plan.RelTraitSet;
+import org.apache.calcite.rel.RelCollationTraitDef;
+import org.apache.calcite.rel.RelInput;
+import org.apache.calcite.rel.RelNode;
+import org.apache.calcite.rel.RelWriter;
+import org.apache.calcite.rel.core.TableScan;
+import org.apache.calcite.rel.hint.RelHint;
+import org.apache.calcite.schema.Table;
+
+import java.util.List;
+
+/**
+ * A LogicalTableScan reads all the rows from a {@link 
RelOptTable}.
+ *
+ * This class is copied from Calcite because the {@link #explainTerms} 
should consider hints.
+ *
+ * If the table is a net.sf.saffron.ext.JdbcTable, then this 
is literally possible.
+ * But for other kinds of tables, there may be many ways to read the data from 
the table. For some
+ * kinds of table, it may not even be possible to read all of the rows unless 
some narrowing
+ * constraint is applied.
+ *
+ * In the example of the net.sf.saffron.ext.ReflectSchema 
schema,
+ *
+ * 
+ *
+ * select from fields
+ *
+ * 
+ *
+ * cannot be implemented, but
+ *
+ * 
+ *
+ * select from fields as f
+ * where f.getClass().getName().equals("java.lang.String")
+ *
+ * 
+ *
+ * can. It is the optimizer's responsibility to find these ways, by 
applying transformation
+ * rules.
+ *
+ * CALCITE-4581, Line 106: {@link #explainTerms} method should consider 
hints.

Review comment:
   Let's use the new strategy to trace modification? i.e. `// BEGIN FLINK 
MODIFICATION`




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-3387) Replace all ByteArrayOutputStream and DataOutputStream use in State Backends by DataOutputSerializer

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-3387.
---
Resolution: Abandoned

Closing as abandoned due to inactivity.

> Replace all ByteArrayOutputStream and DataOutputStream use in State Backends 
> by DataOutputSerializer
> 
>
> Key: FLINK-3387
> URL: https://issues.apache.org/jira/browse/FLINK-3387
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.0.0
>Reporter: Stephan Ewen
>Priority: Major
>
> Currently, the {{byte[]}} to object conversion and vice versa is always done 
> by creating new byte array streams and Data Input/Output streams. That is 
> pretty wasteful, we can optimize this with Flink's own classes:
>   - {{DataInputDeserializer}}
>   - {{DataOutputSerializer}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-3387) Replace all ByteArrayOutputStream and DataOutputStream use in State Backends by DataOutputSerializer

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-3387:

Labels:   (was: stale-assigned)

> Replace all ByteArrayOutputStream and DataOutputStream use in State Backends 
> by DataOutputSerializer
> 
>
> Key: FLINK-3387
> URL: https://issues.apache.org/jira/browse/FLINK-3387
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.0.0
>Reporter: Stephan Ewen
>Priority: Major
>
> Currently, the {{byte[]}} to object conversion and vice versa is always done 
> by creating new byte array streams and Data Input/Output streams. That is 
> pretty wasteful, we can optimize this with Flink's own classes:
>   - {{DataInputDeserializer}}
>   - {{DataOutputSerializer}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-3387) Replace all ByteArrayOutputStream and DataOutputStream use in State Backends by DataOutputSerializer

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-3387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen reassigned FLINK-3387:
---

Assignee: (was: Subhobrata Dey)

> Replace all ByteArrayOutputStream and DataOutputStream use in State Backends 
> by DataOutputSerializer
> 
>
> Key: FLINK-3387
> URL: https://issues.apache.org/jira/browse/FLINK-3387
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.0.0
>Reporter: Stephan Ewen
>Priority: Major
>  Labels: stale-assigned
>
> Currently, the {{byte[]}} to object conversion and vice versa is always done 
> by creating new byte array streams and Data Input/Output streams. That is 
> pretty wasteful, we can optimize this with Flink's own classes:
>   - {{DataInputDeserializer}}
>   - {{DataOutputSerializer}}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-1778) Improve normalized keys in composite key case

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-1778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-1778.
---
Resolution: Abandoned

Closing as abandoned.

We rely less on normalized keys now during sorting, but rather just on the byte 
string representation.

> Improve normalized keys in composite key case
> -
>
> Key: FLINK-1778
> URL: https://issues.apache.org/jira/browse/FLINK-1778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Task
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>Priority: Major
>
> Currently, if we have a key (String, long), the String will take up the 
> entire normalized key space, without being fully discerning anyways. Limiting 
> the key prefix in size and giving space to the second key field should in 
> most cases improve the comparison efficiency.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-4374) GroupReduce Broken for null Date

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-4374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-4374.
---
Resolution: Abandoned

Closing as abandoned, due to long inactivity.

> GroupReduce Broken for null Date
> 
>
> Key: FLINK-4374
> URL: https://issues.apache.org/jira/browse/FLINK-4374
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataSet
>Reporter: Stefan Richter
>Assignee: Timo Walther
>Priority: Major
>  Labels: stale-assigned
>
> The GroupReduceITCase has an error that allows a problem with {{null}} Dates 
> to go uncovered:
>  If I set the parallelism to 1 in {{testDateNullException()}} and all keys 
> actually end up on the same operator, then there is a problem in the 
> de/serialization.
> It seems that {{null}} values are somehow skipped by the serialization 
> process (e.g. maybe no {{null}} indicator is written), which leads to wrong 
> deserializations.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-1778) Improve normalized keys in composite key case

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-1778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-1778:

Labels:   (was: stale-assigned)

> Improve normalized keys in composite key case
> -
>
> Key: FLINK-1778
> URL: https://issues.apache.org/jira/browse/FLINK-1778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Task
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>Priority: Major
>
> Currently, if we have a key (String, long), the String will take up the 
> entire normalized key space, without being fully discerning anyways. Limiting 
> the key prefix in size and giving space to the second key field should in 
> most cases improve the comparison efficiency.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-1778) Improve normalized keys in composite key case

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-1778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen reassigned FLINK-1778:
---

Assignee: (was: Stephan Ewen)

> Improve normalized keys in composite key case
> -
>
> Key: FLINK-1778
> URL: https://issues.apache.org/jira/browse/FLINK-1778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Task
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>Priority: Major
>  Labels: stale-assigned
>
> Currently, if we have a key (String, long), the String will take up the 
> entire normalized key space, without being fully discerning anyways. Limiting 
> the key prefix in size and giving space to the second key field should in 
> most cases improve the comparison efficiency.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] birbilis commented on pull request #227: Update java.md - fixed code typo

2021-04-17 Thread GitBox


birbilis commented on pull request #227:
URL: https://github.com/apache/flink-statefun/pull/227#issuecomment-821814938


   already proposed by other, see mention above


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] birbilis closed pull request #227: Update java.md - fixed code typo

2021-04-17 Thread GitBox


birbilis closed pull request #227:
URL: https://github.com/apache/flink-statefun/pull/227


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-9043) Introduce a friendly way to resume the job from externalized checkpoints automatically

2021-04-17 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-9043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17324255#comment-17324255
 ] 

Stephan Ewen commented on FLINK-9043:
-

I think this might be a good one to revisit now.

The last commonly stored checkpoint storage (S3) is not consistent, so an 
variant that scans the checkpoint directory for the latest checkpoint is 
feasible now. For that approach we would need to resolve FLINK-22008 first.


> Introduce a friendly way to resume the job from externalized checkpoints 
> automatically
> --
>
> Key: FLINK-9043
> URL: https://issues.apache.org/jira/browse/FLINK-9043
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Godfrey He
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> I know a flink job can reovery from checkpoint with restart strategy, but can 
> not recovery as spark streaming jobs when job is starting.
> Every time, the submitted flink job is regarded as a new job, while , in the 
> spark streaming  job, which can detect the checkpoint directory first,  and 
> then recovery from the latest succeed one. However, Flink only can recovery 
> until the job failed first, then retry with strategy.
>  
> So, would flink support to recover from the checkpoint directly in a new job?
> h2. New description by [~sihuazhou]
> Currently, it's quite a bit not friendly for users to recover job from the 
> externalized checkpoint, user need to find the dedicate dir for the job which 
> is not a easy thing when there are too many jobs. This ticket attend to 
> introduce a more friendly way to allow the user to use the externalized 
> checkpoint to do recovery.
> The implementation steps are copied from the comments of [~StephanEwen]:
>  - We could make this an option where you pass a flag (-r) to automatically 
> look for the latest checkpoint in a given directory.
>  - If more than one jobs checkpointed there before, this operation would fail.
>  - We might also need a way to have jobs not create the UUID subdirectory, 
> otherwise the scanning for the latest checkpoint would not easily work.
>   



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] birbilis commented on pull request #224: [hotfix][docs] Fix typos in java.md

2021-04-17 Thread GitBox


birbilis commented on pull request #224:
URL: https://github.com/apache/flink-statefun/pull/224#issuecomment-821814838


   seems I suggested the same fix here - 
https://github.com/apache/flink-statefun/pull/227 - for the 2nd typo, closing 
mine


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] birbilis commented on pull request #228: Update java.md

2021-04-17 Thread GitBox


birbilis commented on pull request #228:
URL: https://github.com/apache/flink-statefun/pull/228#issuecomment-821814585


   not sure what the second change at the "keep learning" shown in the diff is 
about, used GitHub's online editor for the patch


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] birbilis commented on pull request #227: Update java.md - fixed code typo

2021-04-17 Thread GitBox


birbilis commented on pull request #227:
URL: https://github.com/apache/flink-statefun/pull/227#issuecomment-821814574


   not sure what the second change at the "keep learning" shown in the diff is 
about, used GitHub's online editor for the patch


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] birbilis commented on pull request #226: Update java.md

2021-04-17 Thread GitBox


birbilis commented on pull request #226:
URL: https://github.com/apache/flink-statefun/pull/226#issuecomment-821814553


   not sure what the second change at the "keep learning" shown in the diff is 
about, used GitHub's online editor for the patch


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-9043) Introduce a friendly way to resume the job from externalized checkpoints automatically

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-9043:

Labels: pull-request-available  (was: pull-request-available stale-assigned)

> Introduce a friendly way to resume the job from externalized checkpoints 
> automatically
> --
>
> Key: FLINK-9043
> URL: https://issues.apache.org/jira/browse/FLINK-9043
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Godfrey He
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> I know a flink job can reovery from checkpoint with restart strategy, but can 
> not recovery as spark streaming jobs when job is starting.
> Every time, the submitted flink job is regarded as a new job, while , in the 
> spark streaming  job, which can detect the checkpoint directory first,  and 
> then recovery from the latest succeed one. However, Flink only can recovery 
> until the job failed first, then retry with strategy.
>  
> So, would flink support to recover from the checkpoint directly in a new job?
> h2. New description by [~sihuazhou]
> Currently, it's quite a bit not friendly for users to recover job from the 
> externalized checkpoint, user need to find the dedicate dir for the job which 
> is not a easy thing when there are too many jobs. This ticket attend to 
> introduce a more friendly way to allow the user to use the externalized 
> checkpoint to do recovery.
> The implementation steps are copied from the comments of [~StephanEwen]:
>  - We could make this an option where you pass a flag (-r) to automatically 
> look for the latest checkpoint in a given directory.
>  - If more than one jobs checkpointed there before, this operation would fail.
>  - We might also need a way to have jobs not create the UUID subdirectory, 
> otherwise the scanning for the latest checkpoint would not easily work.
>   



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-3389) Add Pre-defined Options settings for RocksDB State backend

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-3389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-3389.
---
Resolution: Done

> Add Pre-defined Options settings for RocksDB State backend
> --
>
> Key: FLINK-3389
> URL: https://issues.apache.org/jira/browse/FLINK-3389
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.0.0
>Reporter: Stephan Ewen
>Assignee: Stephan Ewen
>Priority: Major
>  Labels: stale-assigned
>
> The RocksDB State Backend performance can be optimized for certain settings 
> (for example running on disk or SSD) with certain options.
> Since it is hard to tune for users, we should add a set of predefined options 
> for certain settings.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-9043) Introduce a friendly way to resume the job from externalized checkpoints automatically

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-9043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen reassigned FLINK-9043:
---

Assignee: (was: Sihua Zhou)

> Introduce a friendly way to resume the job from externalized checkpoints 
> automatically
> --
>
> Key: FLINK-9043
> URL: https://issues.apache.org/jira/browse/FLINK-9043
> Project: Flink
>  Issue Type: New Feature
>  Components: Runtime / Checkpointing
>Reporter: Godfrey He
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> I know a flink job can reovery from checkpoint with restart strategy, but can 
> not recovery as spark streaming jobs when job is starting.
> Every time, the submitted flink job is regarded as a new job, while , in the 
> spark streaming  job, which can detect the checkpoint directory first,  and 
> then recovery from the latest succeed one. However, Flink only can recovery 
> until the job failed first, then retry with strategy.
>  
> So, would flink support to recover from the checkpoint directly in a new job?
> h2. New description by [~sihuazhou]
> Currently, it's quite a bit not friendly for users to recover job from the 
> externalized checkpoint, user need to find the dedicate dir for the job which 
> is not a easy thing when there are too many jobs. This ticket attend to 
> introduce a more friendly way to allow the user to use the externalized 
> checkpoint to do recovery.
> The implementation steps are copied from the comments of [~StephanEwen]:
>  - We could make this an option where you pass a flag (-r) to automatically 
> look for the latest checkpoint in a given directory.
>  - If more than one jobs checkpointed there before, this operation would fail.
>  - We might also need a way to have jobs not create the UUID subdirectory, 
> otherwise the scanning for the latest checkpoint would not easily work.
>   



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] birbilis opened a new pull request #228: Update java.md

2021-04-17 Thread GitBox


birbilis opened a new pull request #228:
URL: https://github.com/apache/flink-statefun/pull/228


   typo in json declaration


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] birbilis opened a new pull request #227: Update java.md - fixed code typo

2021-04-17 Thread GitBox


birbilis opened a new pull request #227:
URL: https://github.com/apache/flink-statefun/pull/227


   fixed code snippet typo


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19441) Performance regression on 24.09.2020

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-19441:
-
Labels: pull-request-available  (was: pull-request-available stale-assigned)

> Performance regression on 24.09.2020
> 
>
> Key: FLINK-19441
> URL: https://issues.apache.org/jira/browse/FLINK-19441
> Project: Flink
>  Issue Type: Bug
>Reporter: Arvid Heise
>Assignee: Stephan Ewen
>Priority: Critical
>  Labels: pull-request-available
>
> A couple of benchmarks are showing a small performance regression on 
> 24.09.2020:
> http://codespeed.dak8s.net:8000/timeline/?ben=globalWindow=2
> http://codespeed.dak8s.net:8000/timeline/?ben=tupleKeyBy=2 (?)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-19441) Performance regression on 24.09.2020

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen reassigned FLINK-19441:


Assignee: (was: Stephan Ewen)

> Performance regression on 24.09.2020
> 
>
> Key: FLINK-19441
> URL: https://issues.apache.org/jira/browse/FLINK-19441
> Project: Flink
>  Issue Type: Bug
>Reporter: Arvid Heise
>Priority: Critical
>  Labels: pull-request-available
>
> A couple of benchmarks are showing a small performance regression on 
> 24.09.2020:
> http://codespeed.dak8s.net:8000/timeline/?ben=globalWindow=2
> http://codespeed.dak8s.net:8000/timeline/?ben=tupleKeyBy=2 (?)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19499) Expose Metric Groups to Split Assigners

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-19499:
-
Labels:   (was: stale-assigned)

> Expose Metric Groups to Split Assigners
> ---
>
> Key: FLINK-19499
> URL: https://issues.apache.org/jira/browse/FLINK-19499
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem
>Reporter: Stephan Ewen
>Priority: Major
> Fix For: 1.13.0
>
>
> Split Assigners should have access to metric groups, so they can report 
> metrics on assignment, like pending splits, local-, and remote assignments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-19499) Expose Metric Groups to Split Assigners

2021-04-17 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19499?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen reassigned FLINK-19499:


Assignee: (was: Stephan Ewen)

> Expose Metric Groups to Split Assigners
> ---
>
> Key: FLINK-19499
> URL: https://issues.apache.org/jira/browse/FLINK-19499
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem
>Reporter: Stephan Ewen
>Priority: Major
>  Labels: stale-assigned
> Fix For: 1.13.0
>
>
> Split Assigners should have access to metric groups, so they can report 
> metrics on assignment, like pending splits, local-, and remote assignments.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   >