[jira] [Comment Edited] (FLINK-15571) Create a Redis Streams Connector for Flink

2020-01-13 Thread Yun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014238#comment-17014238
 ] 

Yun Tang edited comment on FLINK-15571 at 1/13/20 11:39 AM:


[~tgrall] , Flink once had a Redis sink connector and moved to [Apache 
Bahir|https://bahir.apache.org/] three years ago, you can refer to the 
discussion mail 
[here|http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Move-Redis-and-Flume-connectors-to-Apache-Bahir-and-redirect-contributions-there-td13102.html]
 .

Since Redis stream is introduced since Redis-5.0 and different from previous 
Redis sink. I think you could start a discussion in dev mail list to talk about 
whether to introduce a new Redis connector.


was (Author: yunta):
Flink once had a Redis sink connector and moved to [Apache 
Bahir|https://bahir.apache.org/] three years ago, you can refer to the 
discussion mail 
[here|http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Move-Redis-and-Flume-connectors-to-Apache-Bahir-and-redirect-contributions-there-td13102.html]
 .

Since Redis stream is introduced since Redis-5.0 and different from previous 
Redis sink. I think you could start a discussion in dev mail list to talk about 
whether to introduce a new Redis connector.

> Create a Redis Streams Connector for Flink
> --
>
> Key: FLINK-15571
> URL: https://issues.apache.org/jira/browse/FLINK-15571
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common
>Reporter: Tugdual Grall
>Priority: Minor
>
> Redis has a "log data structure" called Redis Streams, it would be nice to 
> integrate Redis Streams and Apache Flink as:
>  * Source
>  * Sink
> See Redis Streams introduction: [https://redis.io/topics/streams-intro]
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #10665: [FLINK-15354] Start and stop minikube only in kubernetes related e2e tests

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10665: [FLINK-15354] Start and stop 
minikube only in kubernetes related e2e tests
URL: https://github.com/apache/flink/pull/10665#issuecomment-568440738
 
 
   
   ## CI report:
   
   * 5a0d5d3d499347ca216e19175ff5f066a6d9b458 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142099952) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3861)
 
   * 255d89be8069b36be2b980ea6dba4798568160bb Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/143507134) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4176)
 
   * 9168102e928bacaa8026407f77a33b80a8ddeae4 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/143514708) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4179)
 
   * 1465ba2465b0fba3c91377735481db17abb1c21f Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143667456) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4203)
 
   * 76f3dadc5062279dda9df063901a548315a8b8da Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143714109) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4219)
 
   * 677c4fcd1b82c766f79e296883f8464daafb6a44 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144091754) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4278)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15571) Create a Redis Streams Connector for Flink

2020-01-13 Thread Yun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014238#comment-17014238
 ] 

Yun Tang commented on FLINK-15571:
--

Flink once had a Redis sink connector and moved to [Apache 
Bahir|https://bahir.apache.org/] three years ago, you can refer to the 
discussion mail 
[here|http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Move-Redis-and-Flume-connectors-to-Apache-Bahir-and-redirect-contributions-there-td13102.html]
 .

Since Redis stream is introduced since Redis-5.0 and different from previous 
Redis sink. I think you could start a discussion in dev mail list to talk about 
whether to introduce a new Redis connector.

> Create a Redis Streams Connector for Flink
> --
>
> Key: FLINK-15571
> URL: https://issues.apache.org/jira/browse/FLINK-15571
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common
>Reporter: Tugdual Grall
>Priority: Minor
>
> Redis has a "log data structure" called Redis Streams, it would be nice to 
> integrate Redis Streams and Apache Flink as:
>  * Source
>  * Sink
> See Redis Streams introduction: [https://redis.io/topics/streams-intro]
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #10682: [FLINK-15247][Runtime] Wait for all slots to be free before task executor services shutdown upon stopping

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10682: [FLINK-15247][Runtime] Wait for all 
slots to be free before task executor services shutdown upon stopping
URL: https://github.com/apache/flink/pull/10682#issuecomment-568850864
 
 
   
   ## CI report:
   
   * a7411a5d1a86381078670898c1f20be4fe3a88f4 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142305023) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3903)
 
   * 178357b40789d132e1db954a199bf1c62efff474 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/143770612) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4240)
 
   * 8d2db0ad7b2ce3c268cd8c52a6332a2bc86aca61 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144064773) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4271)
 
   * 5625904d8f9e74bb7a7f1ff392094a8c1ea9b9c2 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/144068497) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4272)
 
   * e3be286cd26d25b91e5805eed149bdf5a39aa292 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144075835) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4275)
 
   * d9b76b1b111564fa5aed4a37888f0ec1001f895d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10674: [FLINK-15220][Connector/Kafka][Table] Add startFromTimestamp in KafkaTableSource

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10674: 
[FLINK-15220][Connector/Kafka][Table] Add startFromTimestamp in KafkaTableSource
URL: https://github.com/apache/flink/pull/10674#issuecomment-568687684
 
 
   
   ## CI report:
   
   * f59106d121f9dd0fc40b23640ec2ad8a663d6020 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142201948) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3880)
 
   * a69cf4921d9a947d437a77c173ae979998ad14a1 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142214808) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3889)
 
   * a8681f8b91fd819cbee65212479e1263fb88bf64 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144107176) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4283)
 
   * 7a75e1c85bc511af3be0eaf977288af816a12b78 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144117224) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4288)
 
   * 632cf6bd9e8f24e2af1da1688227ca3e73fe6858 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] Li-Aihua closed pull request #10409: [FLINK-14918] [flink-end-to-end-perf-tests] Add performance e2e test module and scripts

2020-01-13 Thread GitBox
Li-Aihua closed pull request #10409: [FLINK-14918] 
[flink-end-to-end-perf-tests] Add performance e2e test module and scripts
URL: https://github.com/apache/flink/pull/10409
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15489) Cannot update jobmanager/taskmanager logs

2020-01-13 Thread Yadong Xie (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yadong Xie updated FLINK-15489:
---
Attachment: repro.gif

> Cannot update jobmanager/taskmanager logs
> -
>
> Key: FLINK-15489
> URL: https://issues.apache.org/jira/browse/FLINK-15489
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend
>Affects Versions: 1.9.1, 1.10.0
>Reporter: Dawid Wysakowicz
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
> Attachments: 6E9A4D72-6664-4B00-B02F-57EB2695D967.png, 
> 7A1CF604-032D-46D8-AC70-FBB884E868B9.png, repro.gif, screenshot-1.png, 
> screenshot-2.png
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> There is no way to query the latest state of logs of jobmanager/taskmanager.
> The Web UI show only the first version that was ever displayed.
> How to reproduce:
> * (not sure if necessary) configure logback as described here: 
> https://ci.apache.org/projects/flink/flink-docs-stable/dev/best_practices.html#use-logback-when-running-flink-on-a-cluster
> * start a cluster
> * show jobmanager logs in the Web UI
> * run example job
> * check again the jobmanager logs, there is no trace of the job. Clicking the 
> refresh button does not help



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15573) Let Flink SQL PlannerExpressionParserImpl#FieldRefrence use Unicode as its default charset

2020-01-13 Thread Lsw_aka_laplace (Jira)
Lsw_aka_laplace created FLINK-15573:
---

 Summary: Let Flink SQL PlannerExpressionParserImpl#FieldRefrence 
use Unicode  as its default charset  
 Key: FLINK-15573
 URL: https://issues.apache.org/jira/browse/FLINK-15573
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / Planner
Reporter: Lsw_aka_laplace


Now I am talking about the `PlannerExpressionParserImpl`

    For now  the fieldRefrence‘s  charset is JavaIdentifier,why not change it 
to UnicodeIdentifier?

    Currently in my team, we do actually have this problem. For instance, data 
from Es always contains `@timestamp` field , which can not meet JavaIdentifier. 
So what we did is just let the fieldRefrence Charset use Unicode

 
{code:scala}
 lazy val extensionIdent: Parser[String] = ( "" ~> // handle whitespace 
rep1(acceptIf(Character.isUnicodeIdentifierStart)("identifier expected but '" + 
_ + "' found"), elem("identifier part", Character.isUnicodeIdentifierPart(: 
Char))) ^^ (.mkString) ) 
 lazy val fieldReference: PackratParser[UnresolvedReferenceExpression] = (STAR 
| ident | extensionIdent) ^^ { sym => unresolvedRef(sym) }{code}
 

It is simple but really make sense~

Looking forward for any opinion

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #10838: [hotfix][table-common] Simplify formatted exceptions

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10838: [hotfix][table-common] Simplify 
formatted exceptions
URL: https://github.com/apache/flink/pull/10838#issuecomment-573544312
 
 
   
   ## CI report:
   
   * 6d7cd665ed96239643ccbc81bafacc8fd0e12e80 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/144107161) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4282)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10837: [FLINK-15541] Fix unstable case FlinkKinesisConsumerTest

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10837: [FLINK-15541] Fix unstable case 
FlinkKinesisConsumerTest
URL: https://github.com/apache/flink/pull/10837#issuecomment-573544281
 
 
   
   ## CI report:
   
   * 4d045d632eb72a162e5d56aa52a6819042d2f1aa Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144107145) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4281)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10839: [FLINK-15567][table][docs] Add documentation for INSERT statements for Flink SQL

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10839: [FLINK-15567][table][docs] Add 
documentation for INSERT statements for Flink SQL
URL: https://github.com/apache/flink/pull/10839#issuecomment-573563076
 
 
   
   ## CI report:
   
   * c5595ea0c994dcf664cb68ec45a6f500e705a3b7 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144113663) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4285)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10841: [FLINK-15565][table-planner-blink] Fix error conversion of TinyInt, SmallInt literals between Flink and Calcite

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10841: [FLINK-15565][table-planner-blink] 
Fix error conversion of TinyInt,SmallInt literals between Flink and Calcite
URL: https://github.com/apache/flink/pull/10841#issuecomment-573572659
 
 
   
   ## CI report:
   
   * d3abca3b9c9ca0ad030f7e208843502d14ff13a5 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144117186) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4287)
 
   * b5b112c7954d4961cfbc71f9d268a7a2561e6051 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #10842: [FLINK-15489][web]: add cache control no-cache to log api

2020-01-13 Thread GitBox
flinkbot commented on issue #10842: [FLINK-15489][web]: add cache control 
no-cache to log api
URL: https://github.com/apache/flink/pull/10842#issuecomment-573618229
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit fbc3d61dc5657907e9e55497cf70431ceff2d2d9 (Mon Jan 13 
11:25:21 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
* **This pull request references an unassigned [Jira 
ticket](https://issues.apache.org/jira/browse/FLINK-15489).** According to the 
[code contribution 
guide](https://flink.apache.org/contributing/contribute-code.html), tickets 
need to be assigned before starting with the implementation work.
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] Type of keys should be `BinaryRow`…

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] 
Type of keys should be `BinaryRow`…
URL: https://github.com/apache/flink/pull/10815#issuecomment-572566376
 
 
   
   ## CI report:
   
   * 19a4290f709495491fe460037c8c31d106984ea8 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143732723) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4229)
 
   * c3ef5ea345a343170806de8112163edb7df31f69 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/144110200) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4284)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15489) Cannot update jobmanager/taskmanager logs

2020-01-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15489:
---
Labels: pull-request-available  (was: )

> Cannot update jobmanager/taskmanager logs
> -
>
> Key: FLINK-15489
> URL: https://issues.apache.org/jira/browse/FLINK-15489
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend
>Affects Versions: 1.9.1, 1.10.0
>Reporter: Dawid Wysakowicz
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
> Attachments: 6E9A4D72-6664-4B00-B02F-57EB2695D967.png, 
> 7A1CF604-032D-46D8-AC70-FBB884E868B9.png, screenshot-1.png, screenshot-2.png
>
>
> There is no way to query the latest state of logs of jobmanager/taskmanager.
> The Web UI show only the first version that was ever displayed.
> How to reproduce:
> * (not sure if necessary) configure logback as described here: 
> https://ci.apache.org/projects/flink/flink-docs-stable/dev/best_practices.html#use-logback-when-running-flink-on-a-cluster
> * start a cluster
> * show jobmanager logs in the Web UI
> * run example job
> * check again the jobmanager logs, there is no trace of the job. Clicking the 
> refresh button does not help



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] vthinkxie opened a new pull request #10842: [FLINK-15489][web]: add cache control no-cache to log api

2020-01-13 Thread GitBox
vthinkxie opened a new pull request #10842: [FLINK-15489][web]: add cache 
control no-cache to log api
URL: https://github.com/apache/flink/pull/10842
 
 
   ## What is the purpose of the change
   
   fix https://issues.apache.org/jira/browse/FLINK-14270
   
   ## Brief change log
   
   support more metric display at once
   
   ## Verifying this change
   
 - *Go to the log page*
 - *Click the refresh button*
 - *Check the network panel if the log with no-cache header*
   
   
   before:
   https://user-images.githubusercontent.com/1506722/72252288-f7674f00-3639-11ea-9861-93157e789082.png";>
   
   
   after:
   https://user-images.githubusercontent.com/1506722/72252293-fa623f80-3639-11ea-8a8a-049bf09b2677.png";>
   
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur opened a new pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur opened a new pull request #10832: [FLINK-14163][runtime]Enforce 
synchronous registration of Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832
 
 
   ## What is the purpose of the change
   
   `Execution#producedPartitions` are registered through an asynchronous 
interface `ShuffleMaster#registerPartitionWithProducer` to `ShuffleMaster`, 
however they are not always accessed through callbacks. So, it is possible that 
`Execution#producedPartitions` have not been available yet when accessed (in 
`Execution#deploy` for example). Since the only implementation of 
`ShuffleMaster` is `NettyShuffleMaster`, which indeed registers 
producedPartition in a synchronous way, this pull request enforces synchronous 
registrations under an asynchronous interface as a temporary fix to prevent 
future misuses.
   
   ## Brief change log
   - Make sure the future of registration is done immediately after 
registration. 
   - Throws IllegalStateException otherwise.
   
   ## Verifying this change
   unit test: 
   ExecutionTest#testSynchronousRegisterProducedPartitions
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? JavaDocs
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur removed a comment on issue #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur removed a comment on issue #10832: [FLINK-14163][runtime]Enforce 
synchronous registration of Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#issuecomment-573614132
 
 
   > Thanks for opening the PR @curcur . The change generally looks good to me.
   > I have a few comments for it.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur closed pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur closed pull request #10832: [FLINK-14163][runtime]Enforce synchronous 
registration of Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur commented on issue #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur commented on issue #10832: [FLINK-14163][runtime]Enforce synchronous 
registration of Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#issuecomment-573614132
 
 
   > Thanks for opening the PR @curcur . The change generally looks good to me.
   > I have a few comments for it.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365749861
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionGraphTestUtils.java
 ##
 @@ -431,6 +433,18 @@ public static ExecutionJobVertex 
getExecutionVertex(JobVertexID id) throws Excep
return getExecutionVertex(id, new 
DirectScheduledExecutorService());
}
 
+   public static ExecutionEdge createMockExecutionEdge(int maxParallelism) 
{
 
 Review comment:
   This is reused from ExecutionVertexDeploymentTest.
   
   Let me have a try to see whether I can avoid Mockito and at the same time 
not to make the code too complicated. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] curcur commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365749861
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionGraphTestUtils.java
 ##
 @@ -431,6 +433,18 @@ public static ExecutionJobVertex 
getExecutionVertex(JobVertexID id) throws Excep
return getExecutionVertex(id, new 
DirectScheduledExecutorService());
}
 
+   public static ExecutionEdge createMockExecutionEdge(int maxParallelism) 
{
 
 Review comment:
   This is reused from ExecutionVertexDeploymentTest.
   
   Let me have a try to see whether we can avoid Mockito and at the same time 
not to make the code too complicated. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Comment Edited] (FLINK-14163) Execution#producedPartitions is possibly not assigned when used

2020-01-13 Thread Yuan Mei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014147#comment-17014147
 ] 

Yuan Mei edited comment on FLINK-14163 at 1/13/20 11:04 AM:


 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the point of keeping the interface. 
I think retain a timeout is a good compromise, but may introduce some 
inconsistent system behavior. Hence even if we decide to retain a timeout, I 
would prefer to document as `enforce a synchronous implementation under 
asynchronous interface`.

 


was (Author: ym):
 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the point to keep the interface. I 
think retain a timeout is a good compromise, but may introduce some 
inconsistent system behavior. Hence even if we decide to retain a timeout, I 
would prefer to document as `enforce a synchronous implementation under 
asynchronous interface`.

 

> Execution#producedPartitions is possibly not assigned when used
> ---
>
> Key: FLINK-14163
> URL: https://issues.apache.org/jira/browse/FLINK-14163
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Zhu Zhu
>Assignee: Yuan Mei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently {{Execution#producedPartitions}} is assigned after the partitions 
> have completed the registration to shuffle master in 
> {{Execution#registerProducedPartitions(...)}}.
> The partition registration is an async interface 
> ({{ShuffleMaster#registerPartitionWithProducer(...)}}), so 
> {{Execution#producedPartitions}} is possible[1] not set when used. 
> Usages includes:
> 1. deploying this task, so that the task may be deployed without its result 
> partitions assigned, and the job would hang. (DefaultScheduler issue only, 
> since legacy scheduler handled this case)
> 2. generating input descriptors for downstream tasks: 
> 3. retrieve {{ResultPartitionID}} for partition releasing: 
> [1] If a user uses Flink default shuffle master {{NettyShuffleMaster}}, it is 
> not problematic at the moment since it returns a completed future on 
> registration, so that it would be a synchronized process. However, if users 
> implement their own shuffle service in which the 
> {{ShuffleMaster#registerPartitionWithProducer}} returns an pending future, it 
> can be a problem. This is possible since customizable shuffle service is open 
> to users since 1.9 (via config "shuffle-service-factory.class").
> To avoid issues to happen, we may either 
> 1. fix all the usages of {{Execution#producedPartitions}} regarding the async 
> assigning, or 
> 2. change {{ShuffleMaster#registerPartitionWithProducer(...)}} to a sync 
> interface



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15174) FLINK security using PKI mutual auth needs certificate pinning or Private CA

2020-01-13 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014222#comment-17014222
 ] 

Stephan Ewen commented on FLINK-15174:
--

[~knaufk] We can continue the discussion about whether there are other possible 
workarounds, but the patch as such works and is meaningful in my opinion, so I 
would close this issue.

> FLINK security using PKI mutual auth needs certificate pinning or Private CA
> 
>
> Key: FLINK-15174
> URL: https://issues.apache.org/jira/browse/FLINK-15174
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration, Runtime / REST
>Affects Versions: 1.9.0, 1.9.1, 1.10.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0, 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The current design for Flink security for internal/REST relies on PKI mutual 
> authentication. However, the design is not robust if CA used for generating 
> certificates are public CA or Firwide internal CA. This is due to how the 
> chain of trust works whilst validating the client certificate. i.e. Any 
> certificate signed by same CA would be able to make a connection to internal 
> Flink network.
> Proposed improvement.
> An environment where operators are constrained to use firmwide Internal 
> public CA, Allow the operator to specify the certificate fingerprint to 
> further protect the cluster allowing only specific certificate.
> This change should be a backward compatible change where one can use just 
> certificate with private CA.
> Changes are easy to implement as all network communications are done using 
> netty and netty provides FingerprintTrustManagerFactory.
> Happy to send PR if we agree on the change.
> Document corrections.
> From security documentation.
> [https://ci.apache.org/projects/flink/flink-docs-stable/ops/security-ssl.html]
> _"All internal connections are SSL authenticated and encrypted. The 
> connections use *mutual authentication*, meaning both server and client-side 
> of each connection need to present the certificate to each other. The 
> certificate acts effectively as a shared secret."_
> _-_ This not exactly true. Any party who obtains the client certificate from 
> CA would be able to form the connection even though the certificate 
> public/private keys are different. So it's not *a* shared secret ( merely a 
> common signature)
> _Further doc says - "A common setup is to generate a dedicated certificate 
> (maybe self-signed) for a Flink deployment._
> - I think this is the only way to make the cluster secure. i.e. create 
> private CA just for the cluster.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15174) FLINK security using PKI mutual auth needs certificate pinning or Private CA

2020-01-13 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-15174.


> FLINK security using PKI mutual auth needs certificate pinning or Private CA
> 
>
> Key: FLINK-15174
> URL: https://issues.apache.org/jira/browse/FLINK-15174
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration, Runtime / REST
>Affects Versions: 1.9.0, 1.9.1, 1.10.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0, 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The current design for Flink security for internal/REST relies on PKI mutual 
> authentication. However, the design is not robust if CA used for generating 
> certificates are public CA or Firwide internal CA. This is due to how the 
> chain of trust works whilst validating the client certificate. i.e. Any 
> certificate signed by same CA would be able to make a connection to internal 
> Flink network.
> Proposed improvement.
> An environment where operators are constrained to use firmwide Internal 
> public CA, Allow the operator to specify the certificate fingerprint to 
> further protect the cluster allowing only specific certificate.
> This change should be a backward compatible change where one can use just 
> certificate with private CA.
> Changes are easy to implement as all network communications are done using 
> netty and netty provides FingerprintTrustManagerFactory.
> Happy to send PR if we agree on the change.
> Document corrections.
> From security documentation.
> [https://ci.apache.org/projects/flink/flink-docs-stable/ops/security-ssl.html]
> _"All internal connections are SSL authenticated and encrypted. The 
> connections use *mutual authentication*, meaning both server and client-side 
> of each connection need to present the certificate to each other. The 
> certificate acts effectively as a shared secret."_
> _-_ This not exactly true. Any party who obtains the client certificate from 
> CA would be able to form the connection even though the certificate 
> public/private keys are different. So it's not *a* shared secret ( merely a 
> common signature)
> _Further doc says - "A common setup is to generate a dedicated certificate 
> (maybe self-signed) for a Flink deployment._
> - I think this is the only way to make the cluster secure. i.e. create 
> private CA just for the cluster.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (FLINK-15174) FLINK security using PKI mutual auth needs certificate pinning or Private CA

2020-01-13 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen resolved FLINK-15174.
--
Fix Version/s: (was: 1.9.2)
   1.11.0
   Resolution: Fixed

I resolved this issue for 1.10.0 and 1.11.0


> FLINK security using PKI mutual auth needs certificate pinning or Private CA
> 
>
> Key: FLINK-15174
> URL: https://issues.apache.org/jira/browse/FLINK-15174
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration, Runtime / REST
>Affects Versions: 1.9.0, 1.9.1, 1.10.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0, 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The current design for Flink security for internal/REST relies on PKI mutual 
> authentication. However, the design is not robust if CA used for generating 
> certificates are public CA or Firwide internal CA. This is due to how the 
> chain of trust works whilst validating the client certificate. i.e. Any 
> certificate signed by same CA would be able to make a connection to internal 
> Flink network.
> Proposed improvement.
> An environment where operators are constrained to use firmwide Internal 
> public CA, Allow the operator to specify the certificate fingerprint to 
> further protect the cluster allowing only specific certificate.
> This change should be a backward compatible change where one can use just 
> certificate with private CA.
> Changes are easy to implement as all network communications are done using 
> netty and netty provides FingerprintTrustManagerFactory.
> Happy to send PR if we agree on the change.
> Document corrections.
> From security documentation.
> [https://ci.apache.org/projects/flink/flink-docs-stable/ops/security-ssl.html]
> _"All internal connections are SSL authenticated and encrypted. The 
> connections use *mutual authentication*, meaning both server and client-side 
> of each connection need to present the certificate to each other. The 
> certificate acts effectively as a shared secret."_
> _-_ This not exactly true. Any party who obtains the client certificate from 
> CA would be able to form the connection even though the certificate 
> public/private keys are different. So it's not *a* shared secret ( merely a 
> common signature)
> _Further doc says - "A common setup is to generate a dedicated certificate 
> (maybe self-signed) for a Flink deployment._
> - I think this is the only way to make the cluster secure. i.e. create 
> private CA just for the cluster.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] curcur commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365745754
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionTest.java
 ##
 @@ -540,6 +557,65 @@ public void testSlotReleaseAtomicallyReleasesExecution() 
throws Exception {
});
}
 
+   /**
+* Tests that producedPartitions are registered synchronously under an 
asynchronous interface.
+*/
+   @Test(expected = IllegalStateException.class)
+   public void testSynchronousRegisterProducedPartitions() throws 
Exception {
+   ExecutionJobVertex jobVertex = getExecutionVertex(
+   new JobVertexID(),
+   new DirectScheduledExecutorService(),
+   EAGER);
+   IntermediateResult result =
+   new IntermediateResult(new 
IntermediateDataSetID(), jobVertex, 1, ResultPartitionType.PIPELINED);
+   ExecutionVertex vertex =
+   new ExecutionVertex(jobVertex, 0, new 
IntermediateResult[]{result}, Time.minutes(1));
+   ExecutionEdge edge = createMockExecutionEdge(1);
+   result.getPartitions()[0].addConsumerGroup();
+   result.getPartitions()[0].addConsumer(edge, 0);
+   TaskManagerLocation location =
+   new TaskManagerLocation(ResourceID.generate(), 
InetAddress.getLoopbackAddress(), 1);
+   ExecutionAttemptID attemptID = new ExecutionAttemptID();
+   Execution.registerProducedPartitions(vertex, location, 
attemptID, EAGER.allowLazyDeployment());
+   }
+
+   @Nonnull
+   private ExecutionJobVertex getExecutionVertex(
 
 Review comment:
   that's a good idea, will do it.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15174) FLINK security using PKI mutual auth needs certificate pinning or Private CA

2020-01-13 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014220#comment-17014220
 ] 

Stephan Ewen commented on FLINK-15174:
--

The patch is merged in 

  - 1.11.0 via 16eed2928bad97669e13ad521b2ca9fa5db3c582 and 
69a727d1f9b6ef39ffb707f6d4ef46c534944ab5
  - 1.10.0 via 83db7bc122e48248e8e66b816cbd3ed630878969 and 
7ddd746b2b9d26c02826cc92c58ed9134f34bf0b

> FLINK security using PKI mutual auth needs certificate pinning or Private CA
> 
>
> Key: FLINK-15174
> URL: https://issues.apache.org/jira/browse/FLINK-15174
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration, Runtime / REST
>Affects Versions: 1.9.0, 1.9.1, 1.10.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.2, 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The current design for Flink security for internal/REST relies on PKI mutual 
> authentication. However, the design is not robust if CA used for generating 
> certificates are public CA or Firwide internal CA. This is due to how the 
> chain of trust works whilst validating the client certificate. i.e. Any 
> certificate signed by same CA would be able to make a connection to internal 
> Flink network.
> Proposed improvement.
> An environment where operators are constrained to use firmwide Internal 
> public CA, Allow the operator to specify the certificate fingerprint to 
> further protect the cluster allowing only specific certificate.
> This change should be a backward compatible change where one can use just 
> certificate with private CA.
> Changes are easy to implement as all network communications are done using 
> netty and netty provides FingerprintTrustManagerFactory.
> Happy to send PR if we agree on the change.
> Document corrections.
> From security documentation.
> [https://ci.apache.org/projects/flink/flink-docs-stable/ops/security-ssl.html]
> _"All internal connections are SSL authenticated and encrypted. The 
> connections use *mutual authentication*, meaning both server and client-side 
> of each connection need to present the certificate to each other. The 
> certificate acts effectively as a shared secret."_
> _-_ This not exactly true. Any party who obtains the client certificate from 
> CA would be able to form the connection even though the certificate 
> public/private keys are different. So it's not *a* shared secret ( merely a 
> common signature)
> _Further doc says - "A common setup is to generate a dedicated certificate 
> (maybe self-signed) for a Flink deployment._
> - I think this is the only way to make the cluster secure. i.e. create 
> private CA just for the cluster.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] curcur commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
curcur commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365745182
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionTest.java
 ##
 @@ -540,6 +557,65 @@ public void testSlotReleaseAtomicallyReleasesExecution() 
throws Exception {
});
}
 
+   /**
+* Tests that producedPartitions are registered synchronously under an 
asynchronous interface.
+*/
+   @Test(expected = IllegalStateException.class)
+   public void testSynchronousRegisterProducedPartitions() throws 
Exception {
+   ExecutionJobVertex jobVertex = getExecutionVertex(
+   new JobVertexID(),
+   new DirectScheduledExecutorService(),
+   EAGER);
+   IntermediateResult result =
+   new IntermediateResult(new 
IntermediateDataSetID(), jobVertex, 1, ResultPartitionType.PIPELINED);
+   ExecutionVertex vertex =
+   new ExecutionVertex(jobVertex, 0, new 
IntermediateResult[]{result}, Time.minutes(1));
+   ExecutionEdge edge = createMockExecutionEdge(1);
+   result.getPartitions()[0].addConsumerGroup();
+   result.getPartitions()[0].addConsumer(edge, 0);
 
 Review comment:
   Yep, this part is needed, there has to be one consumer.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15174) FLINK security using PKI mutual auth needs certificate pinning or Private CA

2020-01-13 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014219#comment-17014219
 ] 

Stephan Ewen commented on FLINK-15174:
--

[~knaufk] I tried to describe that above: The limitation in the JDK 
implementation is that any non-self-signed certificate needs the issuing CA in 
the TrustStore, which  would then cause also all other certificates by that CA 
to be trusted. Just having the certificate or even only having the 
purpose-dedicated intermediate CA would not work (validation would fail).

> FLINK security using PKI mutual auth needs certificate pinning or Private CA
> 
>
> Key: FLINK-15174
> URL: https://issues.apache.org/jira/browse/FLINK-15174
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Configuration, Runtime / REST
>Affects Versions: 1.9.0, 1.9.1, 1.10.0
>Reporter: Bhagavan
>Assignee: Bhagavan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.2, 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The current design for Flink security for internal/REST relies on PKI mutual 
> authentication. However, the design is not robust if CA used for generating 
> certificates are public CA or Firwide internal CA. This is due to how the 
> chain of trust works whilst validating the client certificate. i.e. Any 
> certificate signed by same CA would be able to make a connection to internal 
> Flink network.
> Proposed improvement.
> An environment where operators are constrained to use firmwide Internal 
> public CA, Allow the operator to specify the certificate fingerprint to 
> further protect the cluster allowing only specific certificate.
> This change should be a backward compatible change where one can use just 
> certificate with private CA.
> Changes are easy to implement as all network communications are done using 
> netty and netty provides FingerprintTrustManagerFactory.
> Happy to send PR if we agree on the change.
> Document corrections.
> From security documentation.
> [https://ci.apache.org/projects/flink/flink-docs-stable/ops/security-ssl.html]
> _"All internal connections are SSL authenticated and encrypted. The 
> connections use *mutual authentication*, meaning both server and client-side 
> of each connection need to present the certificate to each other. The 
> certificate acts effectively as a shared secret."_
> _-_ This not exactly true. Any party who obtains the client certificate from 
> CA would be able to form the connection even though the certificate 
> public/private keys are different. So it's not *a* shared secret ( merely a 
> common signature)
> _Further doc says - "A common setup is to generate a dedicated certificate 
> (maybe self-signed) for a Flink deployment._
> - I think this is the only way to make the cluster secure. i.e. create 
> private CA just for the cluster.
>  
>  
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-15568) RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 versions

2020-01-13 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15568?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014212#comment-17014212
 ] 

Zhu Zhu edited comment on FLINK-15568 at 1/13/20 10:51 AM:
---

Hi [~andrew_lin], RestartPipelinedRegionStrategy is experimental in 1.8 with 
some known issues. It becomes production ready since 1.9 (FLINK-4256). State 
restore is also supported since then.
There is no plan to backport the changes to 1.8 since the changes are huge and 
not bug fixes.


was (Author: zhuzh):
Hi [~andrew_lin], RestartPipelinedRegionStrategy is experimental in 1.8 with 
some known issues. It becomes production ready since 1.9 (FLINK-4256).
There is no plan to backport the changes to 1.8 since the changes are huge and 
not bug fixes.

> RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 
> versions
> --
>
> Key: FLINK-15568
> URL: https://issues.apache.org/jira/browse/FLINK-15568
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.8.0, 1.8.1, 1.8.3
>Reporter: Andrew.D.lin
>Priority: Minor
> Attachments: image-2020-01-13-16-40-47-888.png
>
>
> In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
> checkpoint and marked TODO.
> Should we support this feature (region restart) in flink 1.8?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15568) RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 versions

2020-01-13 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15568?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014212#comment-17014212
 ] 

Zhu Zhu commented on FLINK-15568:
-

Hi [~andrew_lin], RestartPipelinedRegionStrategy is experimental in 1.8 with 
some known issues. It becomes production ready since 1.9 (FLINK-4256).
There is no plan to backport the changes to 1.8 since the changes are huge and 
not bug fixes.

> RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 
> versions
> --
>
> Key: FLINK-15568
> URL: https://issues.apache.org/jira/browse/FLINK-15568
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.8.0, 1.8.1, 1.8.3
>Reporter: Andrew.D.lin
>Priority: Minor
> Attachments: image-2020-01-13-16-40-47-888.png
>
>
> In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
> checkpoint and marked TODO.
> Should we support this feature (region restart) in flink 1.8?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15066) Cannot run multiple `insert into csvTable values ()`

2020-01-13 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014211#comment-17014211
 ] 

Jingsong Lee commented on FLINK-15066:
--

[~TsReaper] If you are interesting for this, you can take a look to:

https://issues.apache.org/jira/browse/FLINK-14266

https://issues.apache.org/jira/browse/FLINK-14267

Plan to introduce new Csv, and integrated to {{FileSystemOutputFormat}}.

> Cannot run multiple `insert into csvTable values ()`
> 
>
> Key: FLINK-15066
> URL: https://issues.apache.org/jira/browse/FLINK-15066
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Reporter: Kurt Young
>Assignee: Jingsong Lee
>Priority: Major
> Fix For: 1.11.0
>
>
> I created a csv table in sql client, and tried to insert some data into this 
> table.
> The first insert into success, but the second one failed with exception: 
> {code:java}
> // Caused by: java.io.IOException: File or directory /.../xxx.csv already 
> exists. Existing files and directories are not overwritten in NO_OVERWRITE 
> mode. Use OVERWRITE mode to overwrite existing files and directories.at 
> org.apache.flink.core.fs.FileSystem.initOutPathLocalFS(FileSystem.java:817)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15066) Cannot run multiple `insert into csvTable values ()`

2020-01-13 Thread Caizhi Weng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15066?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014210#comment-17014210
 ] 

Caizhi Weng commented on FLINK-15066:
-

>From the perspective of a new user, especially a user from the traditional 
>database, insert overwrite is a really strange behavior. It would be nice if 
>the appending mode is made as the default behavior.

> Cannot run multiple `insert into csvTable values ()`
> 
>
> Key: FLINK-15066
> URL: https://issues.apache.org/jira/browse/FLINK-15066
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Reporter: Kurt Young
>Assignee: Jingsong Lee
>Priority: Major
> Fix For: 1.11.0
>
>
> I created a csv table in sql client, and tried to insert some data into this 
> table.
> The first insert into success, but the second one failed with exception: 
> {code:java}
> // Caused by: java.io.IOException: File or directory /.../xxx.csv already 
> exists. Existing files and directories are not overwritten in NO_OVERWRITE 
> mode. Use OVERWRITE mode to overwrite existing files and directories.at 
> org.apache.flink.core.fs.FileSystem.initOutPathLocalFS(FileSystem.java:817)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-15355) Nightly streaming file sink fails with unshaded hadoop

2020-01-13 Thread Till Rohrmann (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014209#comment-17014209
 ] 

Till Rohrmann edited comment on FLINK-15355 at 1/13/20 10:40 AM:
-

Another instance of the problem: 
https://api.travis-ci.org/v3/job/635957245/log.txt

Full build matrix: https://travis-ci.org/apache/flink/builds/635957218


was (Author: till.rohrmann):
Another instance of the problem: 
https://api.travis-ci.org/v3/job/635957245/log.txt

> Nightly streaming file sink fails with unshaded hadoop
> --
>
> Key: FLINK-15355
> URL: https://issues.apache.org/jira/browse/FLINK-15355
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Arvid Heise
>Assignee: Arvid Heise
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 2h
>  Remaining Estimate: 0h
>
> {code:java}
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>  at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>  at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>  at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:664)
>  at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:213)
>  at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>  at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>  at java.security.AccessController.doPrivileged(Native Method)
>  at javax.security.auth.Subject.doAs(Subject.java:422)
>  at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
>  at 
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>  at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
> Caused by: java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at org.apache.flink.util.ExceptionUtils.rethrow(ExceptionUtils.java:199)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1751)
>  at 
> org.apache.flink.streaming.api.environment.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:94)
>  at 
> org.apache.flink.streaming.api.environment.StreamContextEnvironment.execute(StreamContextEnvironment.java:63)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1628)
>  at StreamingFileSinkProgram.main(StreamingFileSinkProgram.java:77)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498)
>  at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:321)
>  ... 11 more
> Caused by: java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>  at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1746)
>  ... 20 more
> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed to 
> submit JobGraph.
>  at 
> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:326)
>  at 
> java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:870)
>  at 
> java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:852)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:274)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFu

[jira] [Closed] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng closed FLINK-15570.
---
Resolution: Duplicate

> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>  Labels: Usability
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user, especially for users from traditional database, to try out 
> Flink if we support inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15355) Nightly streaming file sink fails with unshaded hadoop

2020-01-13 Thread Till Rohrmann (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014209#comment-17014209
 ] 

Till Rohrmann commented on FLINK-15355:
---

Another instance of the problem: 
https://api.travis-ci.org/v3/job/635957245/log.txt

> Nightly streaming file sink fails with unshaded hadoop
> --
>
> Key: FLINK-15355
> URL: https://issues.apache.org/jira/browse/FLINK-15355
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Arvid Heise
>Assignee: Arvid Heise
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 2h
>  Remaining Estimate: 0h
>
> {code:java}
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error: java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:335)
>  at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:205)
>  at org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:138)
>  at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:664)
>  at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:213)
>  at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>  at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>  at java.security.AccessController.doPrivileged(Native Method)
>  at javax.security.auth.Subject.doAs(Subject.java:422)
>  at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
>  at 
> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>  at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
> Caused by: java.lang.RuntimeException: 
> java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at org.apache.flink.util.ExceptionUtils.rethrow(ExceptionUtils.java:199)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1751)
>  at 
> org.apache.flink.streaming.api.environment.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:94)
>  at 
> org.apache.flink.streaming.api.environment.StreamContextEnvironment.execute(StreamContextEnvironment.java:63)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1628)
>  at StreamingFileSinkProgram.main(StreamingFileSinkProgram.java:77)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498)
>  at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:321)
>  ... 11 more
> Caused by: java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobSubmissionException: Failed to submit 
> JobGraph.
>  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>  at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1895)
>  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1746)
>  ... 20 more
> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed to 
> submit JobGraph.
>  at 
> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:326)
>  at 
> java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:870)
>  at 
> java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:852)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
>  at 
> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:274)
>  at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
>  at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
>  at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
>  at 
> java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:561)
>  at 
> java.util.concurrent.CompletableFuture$Un

[GitHub] [flink] zhuzhurk commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
zhuzhurk commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365732020
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionTest.java
 ##
 @@ -540,6 +557,65 @@ public void testSlotReleaseAtomicallyReleasesExecution() 
throws Exception {
});
}
 
+   /**
+* Tests that producedPartitions are registered synchronously under an 
asynchronous interface.
+*/
+   @Test(expected = IllegalStateException.class)
+   public void testSynchronousRegisterProducedPartitions() throws 
Exception {
+   ExecutionJobVertex jobVertex = getExecutionVertex(
+   new JobVertexID(),
+   new DirectScheduledExecutorService(),
+   EAGER);
+   IntermediateResult result =
+   new IntermediateResult(new 
IntermediateDataSetID(), jobVertex, 1, ResultPartitionType.PIPELINED);
+   ExecutionVertex vertex =
+   new ExecutionVertex(jobVertex, 0, new 
IntermediateResult[]{result}, Time.minutes(1));
+   ExecutionEdge edge = createMockExecutionEdge(1);
+   result.getPartitions()[0].addConsumerGroup();
+   result.getPartitions()[0].addConsumer(edge, 0);
 
 Review comment:
   Are these lines necessary?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15387) Expose missing RocksDB properties out via RocksDBNativeMetricOptions

2020-01-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15387?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15387:
-
Fix Version/s: (was: 1.11.0)

> Expose missing RocksDB properties out via RocksDBNativeMetricOptions
> 
>
> Key: FLINK-15387
> URL: https://issues.apache.org/jira/browse/FLINK-15387
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Yun Tang
>Assignee: Yun Tang
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> When we implements FLINK-15368, we need to expose block cache related metrics 
> of RocksDB out by adding more available options to current 
> RocksDBNativeMetricOptions.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15522) Misleading root cause exception when cancelling the job

2020-01-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15522:
-
Fix Version/s: (was: 1.11.0)

> Misleading root cause exception when cancelling the job
> ---
>
> Key: FLINK-15522
> URL: https://issues.apache.org/jira/browse/FLINK-15522
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission, Runtime / Coordination
>Affects Versions: 1.9.1, 1.10.0, 1.11.0
>Reporter: Till Rohrmann
>Assignee: Zhu Zhu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.9.2, 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> When cancelling a Flink job, the following stack trace gets displayed
> {code}
>  The program finished with the following exception:
> org.apache.flink.client.program.ProgramInvocationException: Job failed. 
> (JobID: d0e8c2026709385166bcc0253c30742e)
> at 
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:262)
> at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:338)
> at 
> org.apache.flink.streaming.api.environment.StreamContextEnvironment.execute(StreamContextEnvironment.java:60)
> at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1507)
> at 
> org.apache.flink.streaming.examples.statemachine.StateMachineExample.main(StateMachineExample.java:142)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:576)
> at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:438)
> at 
> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:274)
> at 
> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:746)
> at 
> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:273)
> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:205)
> at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1010)
> at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1083)
> at 
> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1083)
> Caused by: org.apache.flink.runtime.client.JobCancellationException: Job was 
> cancelled.
> at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:148)
> at 
> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:259)
> ... 18 more
> Caused by: 
> org.apache.flink.runtime.jobmanager.scheduler.NoResourceAvailableException: 
> No pooled slot available and request to ResourceManager for new slot failed
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.slotRequestToResourceManagerFailed(SlotPoolImpl.java:357)
> at 
> org.apache.flink.runtime.jobmaster.slotpool.SlotPoolImpl.lambda$requestSlotFromResourceManager$1(SlotPoolImpl.java:345)
> at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:760)
> at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:736)
> at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:397)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:190)
> at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123)
> at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170)
> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
> at scala.PartialFunction$OrElse.applyOrEls

[GitHub] [flink] zhuzhurk commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
zhuzhurk commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365721185
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/shuffle/ShuffleMaster.java
 ##
 @@ -31,6 +31,10 @@
/**
 * Asynchronously register a partition and its producer with the 
shuffle service.
 *
+* IMPORTANT: Please implement the asynchronous interface in a 
synchronous way.
+* The default scheduler only supports synchronous registration for now.
+* Please refer to {@link 
NettyShuffleMaster#registerPartitionWithProducer} for example.
 
 Review comment:
   I'd prefer to state that "one must return a completed future at the moment". 
It is clear enough so the user do need to refer to NettyShuffleMaster for the 
implementation.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
zhuzhurk commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365720350
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionGraphTestUtils.java
 ##
 @@ -431,6 +433,18 @@ public static ExecutionJobVertex 
getExecutionVertex(JobVertexID id) throws Excep
return getExecutionVertex(id, new 
DirectScheduledExecutorService());
}
 
+   public static ExecutionEdge createMockExecutionEdge(int maxParallelism) 
{
 
 Review comment:
   We are trying to get rid of Mockito in flink tests.
   If we are touching this part, I'd prefer to introduce a mock ExecutionEdge 
class for this purpose. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
zhuzhurk commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365723610
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/Execution.java
 ##
 @@ -605,6 +606,26 @@ public void setInitialState(@Nullable 
JobManagerTaskRestore taskRestore) {
});
}
 
+   /**
+* Register producedPartitions to {@link ShuffleMaster}
+*
+* HACK: Please notice that this method simulates asynchronous 
registration in a synchronous way
+* by making sure the returned {@link CompletableFuture} from {@link 
ShuffleMaster#registerPartitionWithProducer}
+* is done immediately.
+*
+* {@link Execution#producedPartitions} are registered through an 
asynchronous interface
+* {@link ShuffleMaster#registerPartitionWithProducer} to {@link 
ShuffleMaster}, however they are not always
+* accessed through callbacks. So, it is possible that {@link 
Execution#producedPartitions}
+* have not been available yet when accessed (in {@link 
Execution#deploy} for example).
+*
+* Since the only implementation of {@link ShuffleMaster} is {@link 
NettyShuffleMaster},
+* which indeed registers producedPartition in a synchronous way, hence 
this method enforces
+* synchronous registration under an asynchronous interface for now.
+*
+* If asynchronous registration is needed in the future, use 
callbacks to access {@link Execution#producedPartitions}.
 
 Review comment:
   This is a temporary fix. So we'd better have a TODO here as a reminder for 
follow up fixes.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhuzhurk commented on a change in pull request #10832: [FLINK-14163][runtime]Enforce synchronous registration of Execution#producedPartitions

2020-01-13 Thread GitBox
zhuzhurk commented on a change in pull request #10832: 
[FLINK-14163][runtime]Enforce synchronous registration of 
Execution#producedPartitions
URL: https://github.com/apache/flink/pull/10832#discussion_r365729537
 
 

 ##
 File path: 
flink-runtime/src/test/java/org/apache/flink/runtime/executiongraph/ExecutionTest.java
 ##
 @@ -540,6 +557,65 @@ public void testSlotReleaseAtomicallyReleasesExecution() 
throws Exception {
});
}
 
+   /**
+* Tests that producedPartitions are registered synchronously under an 
asynchronous interface.
+*/
+   @Test(expected = IllegalStateException.class)
+   public void testSynchronousRegisterProducedPartitions() throws 
Exception {
+   ExecutionJobVertex jobVertex = getExecutionVertex(
+   new JobVertexID(),
+   new DirectScheduledExecutorService(),
+   EAGER);
+   IntermediateResult result =
+   new IntermediateResult(new 
IntermediateDataSetID(), jobVertex, 1, ResultPartitionType.PIPELINED);
+   ExecutionVertex vertex =
+   new ExecutionVertex(jobVertex, 0, new 
IntermediateResult[]{result}, Time.minutes(1));
+   ExecutionEdge edge = createMockExecutionEdge(1);
+   result.getPartitions()[0].addConsumerGroup();
+   result.getPartitions()[0].addConsumer(edge, 0);
+   TaskManagerLocation location =
+   new TaskManagerLocation(ResourceID.generate(), 
InetAddress.getLoopbackAddress(), 1);
+   ExecutionAttemptID attemptID = new ExecutionAttemptID();
+   Execution.registerProducedPartitions(vertex, location, 
attemptID, EAGER.allowLazyDeployment());
+   }
+
+   @Nonnull
+   private ExecutionJobVertex getExecutionVertex(
 
 Review comment:
   Could we extend ExecutionGraphTestUtils#getExecutionVertex(...) to customize 
shuffleMaster and reuse it?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] twalthr commented on issue #10838: [hotfix][table-common] Simplify formatted exceptions

2020-01-13 Thread GitBox
twalthr commented on issue #10838: [hotfix][table-common] Simplify formatted 
exceptions
URL: https://github.com/apache/flink/pull/10838#issuecomment-573595958
 
 
   I'm fine with a builder approach. This also allows us to include error codes 
and other properties (such as severity) in the future. I will update the PR.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15417) Remove the docker volume or mount when starting Mesos e2e cluster

2020-01-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15417?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-15417:
-
Issue Type: Bug  (was: Test)

> Remove the docker volume or mount when starting Mesos e2e cluster
> -
>
> Key: FLINK-15417
> URL: https://issues.apache.org/jira/browse/FLINK-15417
> Project: Flink
>  Issue Type: Bug
>Reporter: Yangze Guo
>Assignee: Yangze Guo
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> As discussed 
> [here|https://github.com/apache/flink/pull/10695#discussion_r361574394], 
> there  is a potential risk of permission problems when cleanup logs and 
> output. We could found another way to let containers get the input and output 
> file.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Kurt Young (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014190#comment-17014190
 ] 

Kurt Young commented on FLINK-15570:


This is duplicated with FLINK-15066?

> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>  Labels: Usability
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user, especially for users from traditional database, to try out 
> Flink if we support inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young reassigned FLINK-15565:
--

Assignee: Jingsong Lee

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Assignee: Jingsong Lee
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|380|65569|4294967425|0.94|28.93|false|sarah robinson|2013-03-01 
> 09:11:58.703176|88.81|xylophone band
>  66|478|65669|4294967339|23.66|38.3

[jira] [Commented] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread Kurt Young (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014186#comment-17014186
 ] 

Kurt Young commented on FLINK-15565:


[~lzljs3620320] I've assigned to you.

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Assignee: Jingsong Lee
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|380|65569|4294967425|0.94|28.93|false|sarah robinson|2013-03-01 
> 09:11:58.703

[GitHub] [flink] carp84 merged pull request #10805: [FLINK-15529] [docs] Update upgrade compatibility table in operations doc for 1.10.0

2020-01-13 Thread GitBox
carp84 merged pull request #10805: [FLINK-15529] [docs] Update upgrade 
compatibility table in operations doc for 1.10.0
URL: https://github.com/apache/flink/pull/10805
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15572) Function DDL is not compliant with FLIP-64

2020-01-13 Thread Timo Walther (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014182#comment-17014182
 ] 

Timo Walther commented on FLINK-15572:
--

CC [~phoenixjiangnan]

> Function DDL is not compliant with FLIP-64
> --
>
> Key: FLINK-15572
> URL: https://issues.apache.org/jira/browse/FLINK-15572
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Timo Walther
>Priority: Major
>
> The Function DDL does not follow the path resolution defined in FLIP-64. It 
> always assumes fully qualified paths for catalog and database instead of 
> resolving partially defined paths with the help of the current 
> catalog/database. See {{createTemporaryView()}} as an example.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-14163) Execution#producedPartitions is possibly not assigned when used

2020-01-13 Thread Yuan Mei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014147#comment-17014147
 ] 

Yuan Mei edited comment on FLINK-14163 at 1/13/20 9:59 AM:
---

 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the point to keep the interface. I 
think retain a timeout is a good compromise, but may introduce some 
inconsistent system behavior. Hence even if we decide to retain a timeout, I 
would prefer to document as `enforce a synchronous implementation under 
asynchronous interface`.

 


was (Author: ym):
 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than keeping the interface. I think 
retain a timeout is a good compromise, but may introduce some inconsistent 
system behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 

> Execution#producedPartitions is possibly not assigned when used
> ---
>
> Key: FLINK-14163
> URL: https://issues.apache.org/jira/browse/FLINK-14163
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Zhu Zhu
>Assignee: Yuan Mei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently {{Execution#producedPartitions}} is assigned after the partitions 
> have completed the registration to shuffle master in 
> {{Execution#registerProducedPartitions(...)}}.
> The partition registration is an async interface 
> ({{ShuffleMaster#registerPartitionWithProducer(...)}}), so 
> {{Execution#producedPartitions}} is possible[1] not set when used. 
> Usages includes:
> 1. deploying this task, so that the task may be deployed without its result 
> partitions assigned, and the job would hang. (DefaultScheduler issue only, 
> since legacy scheduler handled this case)
> 2. generating input descriptors for downstream tasks: 
> 3. retrieve {{ResultPartitionID}} for partition releasing: 
> [1] If a user uses Flink default shuffle master {{NettyShuffleMaster}}, it is 
> not problematic at the moment since it returns a completed future on 
> registration, so that it would be a synchronized process. However, if users 
> implement their own shuffle service in which the 
> {{ShuffleMaster#registerPartitionWithProducer}} returns an pending future, it 
> can be a problem. This is possible since customizable shuffle service is open 
> to users since 1.9 (via config "shuffle-service-factory.class").
> To avoid issues to happen, we may either 
> 1. fix all the usages of {{Execution#producedPartitions}} regarding the async 
> assigning, or 
> 2. change {{ShuffleMaster#registerPartitionWithProducer(...)}} to a sync 
> interface



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-14163) Execution#producedPartitions is possibly not assigned when used

2020-01-13 Thread Yuan Mei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014147#comment-17014147
 ] 

Yuan Mei edited comment on FLINK-14163 at 1/13/20 9:58 AM:
---

 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than keeping the interface. I think 
retain a timeout is a good compromise, but may introduce some inconsistent 
system behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 


was (Author: ym):
 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the interface. I think retain a 
timeout is a good compromise, but may introduce some inconsistent system 
behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 

> Execution#producedPartitions is possibly not assigned when used
> ---
>
> Key: FLINK-14163
> URL: https://issues.apache.org/jira/browse/FLINK-14163
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Zhu Zhu
>Assignee: Yuan Mei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently {{Execution#producedPartitions}} is assigned after the partitions 
> have completed the registration to shuffle master in 
> {{Execution#registerProducedPartitions(...)}}.
> The partition registration is an async interface 
> ({{ShuffleMaster#registerPartitionWithProducer(...)}}), so 
> {{Execution#producedPartitions}} is possible[1] not set when used. 
> Usages includes:
> 1. deploying this task, so that the task may be deployed without its result 
> partitions assigned, and the job would hang. (DefaultScheduler issue only, 
> since legacy scheduler handled this case)
> 2. generating input descriptors for downstream tasks: 
> 3. retrieve {{ResultPartitionID}} for partition releasing: 
> [1] If a user uses Flink default shuffle master {{NettyShuffleMaster}}, it is 
> not problematic at the moment since it returns a completed future on 
> registration, so that it would be a synchronized process. However, if users 
> implement their own shuffle service in which the 
> {{ShuffleMaster#registerPartitionWithProducer}} returns an pending future, it 
> can be a problem. This is possible since customizable shuffle service is open 
> to users since 1.9 (via config "shuffle-service-factory.class").
> To avoid issues to happen, we may either 
> 1. fix all the usages of {{Execution#producedPartitions}} regarding the async 
> assigning, or 
> 2. change {{ShuffleMaster#registerPartitionWithProducer(...)}} to a sync 
> interface



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15572) Function DDL is not compliant with FLIP-64

2020-01-13 Thread Timo Walther (Jira)
Timo Walther created FLINK-15572:


 Summary: Function DDL is not compliant with FLIP-64
 Key: FLINK-15572
 URL: https://issues.apache.org/jira/browse/FLINK-15572
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Reporter: Timo Walther


The Function DDL does not follow the path resolution defined in FLIP-64. It 
always assumes fully qualified paths for catalog and database instead of 
resolving partially defined paths with the help of the current 
catalog/database. See {{createTemporaryView()}} as an example.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15488) Cannot start a taskmanger if using logback

2020-01-13 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao closed FLINK-15488.

Resolution: Fixed

1.10: 6c886ade7c02840423552025f5d5da535ae40d02
master: d76f2166396e3758c08851d024e2c23c8a297062

> Cannot start a taskmanger if using logback
> --
>
> Key: FLINK-15488
> URL: https://issues.apache.org/jira/browse/FLINK-15488
> Project: Flink
>  Issue Type: Bug
>  Components: API / Core, Deployment / Scripts
>Affects Versions: 1.10.0
>Reporter: Dawid Wysakowicz
>Assignee: Yangze Guo
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> When using logback it is not possible to start the taskmanager using 
> {{taskamanger.sh}} scripts. The same problem (probably) occurs when using 
> slf4j that logs into the console.
> The problem is that when calculating memory configuration with 
> {{BashJavaUtils}} class the result is returned through standard output. If 
> something is logged into the console it may result in undefined behavior such 
> as e.g. 
> {code}
> Error: Could not find or load main class 13:51:23.961
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] GJL closed pull request #10804: [FLINK-15488] Obtain the JVM and TM param correctly

2020-01-13 Thread GitBox
GJL closed pull request #10804: [FLINK-15488] Obtain the JVM and TM param 
correctly
URL: https://github.com/apache/flink/pull/10804
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] jrthe42 closed pull request #6666: [FLINK-10290] [table] Fix conversion error in StreamScan and BatchScan

2020-01-13 Thread GitBox
jrthe42 closed pull request #: [FLINK-10290] [table] Fix conversion error 
in StreamScan and BatchScan
URL: https://github.com/apache/flink/pull/
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] jrthe42 closed pull request #6847: [FLINK-10553] [sql] Unified sink and source table name in SQL statement

2020-01-13 Thread GitBox
jrthe42 closed pull request #6847: [FLINK-10553] [sql] Unified sink and source 
table name in SQL statement
URL: https://github.com/apache/flink/pull/6847
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] jrthe42 closed pull request #6301: [FLINK-9794] [jdbc] JDBCOutputFormat does not consider idle connection and multithreads synchronization

2020-01-13 Thread GitBox
jrthe42 closed pull request #6301: [FLINK-9794] [jdbc] JDBCOutputFormat does 
not consider idle connection and multithreads synchronization
URL: https://github.com/apache/flink/pull/6301
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10674: [FLINK-15220][Connector/Kafka][Table] Add startFromTimestamp in KafkaTableSource

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10674: 
[FLINK-15220][Connector/Kafka][Table] Add startFromTimestamp in KafkaTableSource
URL: https://github.com/apache/flink/pull/10674#issuecomment-568687684
 
 
   
   ## CI report:
   
   * f59106d121f9dd0fc40b23640ec2ad8a663d6020 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142201948) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3880)
 
   * a69cf4921d9a947d437a77c173ae979998ad14a1 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142214808) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3889)
 
   * a8681f8b91fd819cbee65212479e1263fb88bf64 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144107176) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4283)
 
   * 7a75e1c85bc511af3be0eaf977288af816a12b78 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10665: [FLINK-15354] Start and stop minikube only in kubernetes related e2e tests

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10665: [FLINK-15354] Start and stop 
minikube only in kubernetes related e2e tests
URL: https://github.com/apache/flink/pull/10665#issuecomment-568440738
 
 
   
   ## CI report:
   
   * 5a0d5d3d499347ca216e19175ff5f066a6d9b458 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/142099952) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=3861)
 
   * 255d89be8069b36be2b980ea6dba4798568160bb Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/143507134) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4176)
 
   * 9168102e928bacaa8026407f77a33b80a8ddeae4 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/143514708) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4179)
 
   * 1465ba2465b0fba3c91377735481db17abb1c21f Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143667456) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4203)
 
   * 76f3dadc5062279dda9df063901a548315a8b8da Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143714109) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4219)
 
   * 677c4fcd1b82c766f79e296883f8464daafb6a44 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144091754) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4278)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-15571) Create a Redis Streams Connector for Flink

2020-01-13 Thread Tugdual Grall (Jira)
Tugdual Grall created FLINK-15571:
-

 Summary: Create a Redis Streams Connector for Flink
 Key: FLINK-15571
 URL: https://issues.apache.org/jira/browse/FLINK-15571
 Project: Flink
  Issue Type: New Feature
  Components: Connectors / Common
Reporter: Tugdual Grall


Redis has a "log data structure" called Redis Streams, it would be nice to 
integrate Redis Streams and Apache Flink as:
 * Source
 * Sink

See Redis Streams introduction: [https://redis.io/topics/streams-intro]

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-15570:

Priority: Major  (was: Minor)

> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Major
>  Labels: Usability
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user, especially for users from traditional database, to try out 
> Flink if we support inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15562) Unable to create walk through project

2020-01-13 Thread Yangze Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15562?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014152#comment-17014152
 ] 

Yangze Guo commented on FLINK-15562:


Hi, [~chesnay], I think it will benefit the user experience if we give a 
suggestion of locking the version of the plugin to 2.4 here. Indeed, some user 
fix the issue in that way in Chinese community. WDYT?

> Unable to create walk through project
> -
>
> Key: FLINK-15562
> URL: https://issues.apache.org/jira/browse/FLINK-15562
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.10.0
>Reporter: Jeff Zhang
>Priority: Major
>
> I try to follow the instruction here to create flink walk though project, but 
> hit the following errors.
> [https://ci.apache.org/projects/flink/flink-docs-master/getting-started/walkthroughs/table_api.html]
> {code:java}
> ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-archetype-plugin:3.0.1:generate (default-cli) 
> on project standalone-pom: archetypeCatalog 
> 'https://repository.apache.org/content/repositories/snapshots/' is not 
> supported anymore. Please read the plugin documentation for details. -> [Help 
> 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15562) Unable to create walk through project

2020-01-13 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15562?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-15562.

Resolution: Not A Problem

The issue is mentioned in the documentation.

> Unable to create walk through project
> -
>
> Key: FLINK-15562
> URL: https://issues.apache.org/jira/browse/FLINK-15562
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.10.0
>Reporter: Jeff Zhang
>Priority: Major
>
> I try to follow the instruction here to create flink walk though project, but 
> hit the following errors.
> [https://ci.apache.org/projects/flink/flink-docs-master/getting-started/walkthroughs/table_api.html]
> {code:java}
> ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-archetype-plugin:3.0.1:generate (default-cli) 
> on project standalone-pom: archetypeCatalog 
> 'https://repository.apache.org/content/repositories/snapshots/' is not 
> supported anymore. Please read the plugin documentation for details. -> [Help 
> 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #10841: [FLINK-15565][table-planner-blink] Fix error conversion of TinyInt, SmallInt literals between Flink and Calcite

2020-01-13 Thread GitBox
flinkbot commented on issue #10841: [FLINK-15565][table-planner-blink] Fix 
error conversion of TinyInt,SmallInt literals between Flink and Calcite
URL: https://github.com/apache/flink/pull/10841#issuecomment-573572659
 
 
   
   ## CI report:
   
   * d3abca3b9c9ca0ad030f7e208843502d14ff13a5 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10839: [FLINK-15567][table][docs] Add documentation for INSERT statements for Flink SQL

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10839: [FLINK-15567][table][docs] Add 
documentation for INSERT statements for Flink SQL
URL: https://github.com/apache/flink/pull/10839#issuecomment-573563076
 
 
   
   ## CI report:
   
   * c5595ea0c994dcf664cb68ec45a6f500e705a3b7 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144113663) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4285)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10840: [FLINK-15444][runtime] Make the component AbstractInvokable in CheckpointBarrierHandler NonNull

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10840: [FLINK-15444][runtime] Make the 
component AbstractInvokable in CheckpointBarrierHandler NonNull
URL: https://github.com/apache/flink/pull/10840#issuecomment-573563119
 
 
   
   ## CI report:
   
   * 61e9e675667e05e44e9c4b5ffd7565f60aec6d62 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144113694) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4286)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10838: [hotfix][table-common] Simplify formatted exceptions

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10838: [hotfix][table-common] Simplify 
formatted exceptions
URL: https://github.com/apache/flink/pull/10838#issuecomment-573544312
 
 
   
   ## CI report:
   
   * 6d7cd665ed96239643ccbc81bafacc8fd0e12e80 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/144107161) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4282)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhijiangW merged pull request #10735: [hotfix][runtime] Fix variable name typos in TaskInformation.java and…

2020-01-13 Thread GitBox
zhijiangW merged pull request #10735: [hotfix][runtime] Fix variable name typos 
in TaskInformation.java and…
URL: https://github.com/apache/flink/pull/10735
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15556) Add a switch for PushProjectIntoTableSourceScanRule

2020-01-13 Thread godfrey he (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014148#comment-17014148
 ] 

godfrey he commented on FLINK-15556:


[~jark], I think if the query only contains one kind of TableSource, the answer 
is yes. But, if there are more than one kind of TableSource in a query, the 
answer may be no. 

> Add a switch for PushProjectIntoTableSourceScanRule
> ---
>
> Key: FLINK-15556
> URL: https://issues.apache.org/jira/browse/FLINK-15556
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: hailong wang
>Priority: Minor
> Fix For: 1.11.0
>
>
> In some case, For the project push down, The  digest of tableSource maybe 
> different. For example, If we create a jdbc tablesource, and is used twice 
> after it for different column, the source can not be reused for the different 
> digest. In this case, the query io will more precious, So we can add a switch 
> to turn off project push down, and the reused the source.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14163) Execution#producedPartitions is possibly not assigned when used

2020-01-13 Thread Yuan Mei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014147#comment-17014147
 ] 

Yuan Mei commented on FLINK-14163:
--

 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the interface. I think retain a 
timeout is a good compromise, but may introduce some inconsistent system 
behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 

> Execution#producedPartitions is possibly not assigned when used
> ---
>
> Key: FLINK-14163
> URL: https://issues.apache.org/jira/browse/FLINK-14163
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Zhu Zhu
>Assignee: Yuan Mei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently {{Execution#producedPartitions}} is assigned after the partitions 
> have completed the registration to shuffle master in 
> {{Execution#registerProducedPartitions(...)}}.
> The partition registration is an async interface 
> ({{ShuffleMaster#registerPartitionWithProducer(...)}}), so 
> {{Execution#producedPartitions}} is possible[1] not set when used. 
> Usages includes:
> 1. deploying this task, so that the task may be deployed without its result 
> partitions assigned, and the job would hang. (DefaultScheduler issue only, 
> since legacy scheduler handled this case)
> 2. generating input descriptors for downstream tasks: 
> 3. retrieve {{ResultPartitionID}} for partition releasing: 
> [1] If a user uses Flink default shuffle master {{NettyShuffleMaster}}, it is 
> not problematic at the moment since it returns a completed future on 
> registration, so that it would be a synchronized process. However, if users 
> implement their own shuffle service in which the 
> {{ShuffleMaster#registerPartitionWithProducer}} returns an pending future, it 
> can be a problem. This is possible since customizable shuffle service is open 
> to users since 1.9 (via config "shuffle-service-factory.class").
> To avoid issues to happen, we may either 
> 1. fix all the usages of {{Execution#producedPartitions}} regarding the async 
> assigning, or 
> 2. change {{ShuffleMaster#registerPartitionWithProducer(...)}} to a sync 
> interface



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-14163) Execution#producedPartitions is possibly not assigned when used

2020-01-13 Thread Yuan Mei (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014147#comment-17014147
 ] 

Yuan Mei edited comment on FLINK-14163 at 1/13/20 9:20 AM:
---

 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the interface. I think retain a 
timeout is a good compromise, but may introduce some inconsistent system 
behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 


was (Author: ym):
 

Synced up with [~zjwang] offline. 

 

In general, I would prefer not to change interfaces unless having to, since 
interface changes have more profound impact on users than implementation 
changes. Especially in this case, given the future design is not determined 
(both Shuffle and Scheduler), changing async -> sycn and then sync -> async 
again bring much more pain to users than keeping the asycn interfaces and 
enforces a sync implementation underneath. 

In the latter case, a user barely needs to do anything if we do want to support 
async implementation in the future, while in the first case, every single user 
that uses the interface has to make a new release to adjust to the updated 
interface.

 

I do not have strong preferences other than the interface. I think retain a 
timeout is a good compromise, but may introduce some inconsistent system 
behavior. Hence even if we decide to retain a timeout, I would prefer to 
document as `enforce a synchronous implementation under asynchronous interface`.

 

> Execution#producedPartitions is possibly not assigned when used
> ---
>
> Key: FLINK-14163
> URL: https://issues.apache.org/jira/browse/FLINK-14163
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.9.0, 1.10.0
>Reporter: Zhu Zhu
>Assignee: Yuan Mei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently {{Execution#producedPartitions}} is assigned after the partitions 
> have completed the registration to shuffle master in 
> {{Execution#registerProducedPartitions(...)}}.
> The partition registration is an async interface 
> ({{ShuffleMaster#registerPartitionWithProducer(...)}}), so 
> {{Execution#producedPartitions}} is possible[1] not set when used. 
> Usages includes:
> 1. deploying this task, so that the task may be deployed without its result 
> partitions assigned, and the job would hang. (DefaultScheduler issue only, 
> since legacy scheduler handled this case)
> 2. generating input descriptors for downstream tasks: 
> 3. retrieve {{ResultPartitionID}} for partition releasing: 
> [1] If a user uses Flink default shuffle master {{NettyShuffleMaster}}, it is 
> not problematic at the moment since it returns a completed future on 
> registration, so that it would be a synchronized process. However, if users 
> implement their own shuffle service in which the 
> {{ShuffleMaster#registerPartitionWithProducer}} returns an pending future, it 
> can be a problem. This is possible since customizable shuffle service is open 
> to users since 1.9 (via config "shuffle-service-factory.class").
> To avoid issues to happen, we may either 
> 1. fix all the usages of {{Execution#producedPartitions}} regarding the async 
> assigning, or 
> 2. change {{ShuffleMaster#registerPartitionWithProducer(...)}} to a sync 
> interface



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-15570:

Description: 
Currently we only support inserting into empty csv tables. It would be nice for 
a new user, especially for users from traditional database, to try out Flink if 
we support inserting into non-empty csv tables.

In other words, it would be nice if the following SQL produces a valid result:
{code:java}
CREATE TABLE myTable(
a INT,
b DOUBLE
) WITH (
'connector.type' = 'filesystem',
'connector.path' = an existing csv file,
'format.type' = 'csv',
'format.derive-schema' = 'true'
);

SELECT * FROM myTable; // produces 3 rows

INSERT INTO myTable VALUES (4, 4.4);

SELECT * FROM myTable; // produces 4 rows{code}
We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
NO_OVERWRITE mode we can only insert into empty csv tables, while in OVERWRITE 
mode inserting into a csv table will wipe all existing data in it (in the above 
example, the last SELECT will produce only 1 instead of 4 rows) which is really 
strange for a mere INSERT operation.

We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
mode to something like appending to files. Supporting specifying write mode in 
the WITH clause is also needed.

  was:
Currently we only support inserting into empty csv tables. It would be nice for 
a new user from traditional database to try out Flink if we support inserting 
into non-empty csv tables.

In other words, it would be nice if the following SQL produces a valid result:
{code:java}
CREATE TABLE myTable(
a INT,
b DOUBLE
) WITH (
'connector.type' = 'filesystem',
'connector.path' = an existing csv file,
'format.type' = 'csv',
'format.derive-schema' = 'true'
);

SELECT * FROM myTable; // produces 3 rows

INSERT INTO myTable VALUES (4, 4.4);

SELECT * FROM myTable; // produces 4 rows{code}
We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
NO_OVERWRITE mode we can only insert into empty csv tables, while in OVERWRITE 
mode inserting into a csv table will wipe all existing data in it (in the above 
example, the last SELECT will produce only 1 instead of 4 rows) which is really 
strange for a mere INSERT operation.

We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
mode to something like appending to files. Supporting specifying write mode in 
the WITH clause is also needed.


> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Minor
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user, especially for users from traditional database, to try out 
> Flink if we support inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-15570:

Labels: Usability  (was: )

> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Minor
>  Labels: Usability
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user, especially for users from traditional database, to try out 
> Flink if we support inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15559) Missing {{site.baseurl}} in some docs

2020-01-13 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014144#comment-17014144
 ] 

Benchao Li commented on FLINK-15559:


The reason is following:

Currently {{build_docs.sh}} does not use {{baseurl}}, for a url like 
{{/dev/table/streaming/query_configuration.html}}, local url is 
{{localhost:4000/dev/table/streaming/query_configuration.html}} which is a 
valid url.

When we has a baseurl, for example {{/projects/flink/flink-docs-master}}, then 
{{localhost:4000/dev/table/streaming/query_configuration.html}} is a broken 
url. In this case, it needs to be 
{{localhost:4000/projects/flink/flink-docs-master/dev/table/streaming/query_configuration.html}}.

 

You can check 
[https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/systemFunctions.html#type-conversion-functions]

The "See the supported types here"'s like is broken due to lacking {{baseurl}}

> Missing {{site.baseurl}} in some docs
> -
>
> Key: FLINK-15559
> URL: https://issues.apache.org/jira/browse/FLINK-15559
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 1.10.0
>Reporter: Benchao Li
>Assignee: Benchao Li
>Priority: Major
>
> For example, in 
>  * dev/table/functions/systemFunctions.md
>  * dev/table/functions/systemFunctions.zh.md
>  * dev/table/sourceSinks.md
>  * dev/table/sourceSinks.zh.md
>  * dev/table/sql/queries.zh.md
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-15559) Missing {{site.baseurl}} in some docs

2020-01-13 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014144#comment-17014144
 ] 

Benchao Li edited comment on FLINK-15559 at 1/13/20 9:19 AM:
-

The reason is following:

Currently {{build_docs.sh}} does not use {{baseurl}}, for a url like 
{{/dev/table/streaming/query_configuration.html}}, local url is 
{{localhost:4000/dev/table/streaming/query_configuration.html}} which is a 
valid url.

When we has a baseurl, for example {{/projects/flink/flink-docs-master}}, then 
{{localhost:4000/dev/table/streaming/query_configuration.html}} is a broken 
url. In this case, it needs to be 
{{localhost:4000/projects/flink/flink-docs-master/dev/table/streaming/query_configuration.html}}.

 

You can check 
[https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/systemFunctions.html#type-conversion-functions]

The "See the supported types here"'s link is broken due to lacking {{baseurl}}


was (Author: libenchao):
The reason is following:

Currently {{build_docs.sh}} does not use {{baseurl}}, for a url like 
{{/dev/table/streaming/query_configuration.html}}, local url is 
{{localhost:4000/dev/table/streaming/query_configuration.html}} which is a 
valid url.

When we has a baseurl, for example {{/projects/flink/flink-docs-master}}, then 
{{localhost:4000/dev/table/streaming/query_configuration.html}} is a broken 
url. In this case, it needs to be 
{{localhost:4000/projects/flink/flink-docs-master/dev/table/streaming/query_configuration.html}}.

 

You can check 
[https://ci.apache.org/projects/flink/flink-docs-master/dev/table/functions/systemFunctions.html#type-conversion-functions]

The "See the supported types here"'s like is broken due to lacking {{baseurl}}

> Missing {{site.baseurl}} in some docs
> -
>
> Key: FLINK-15559
> URL: https://issues.apache.org/jira/browse/FLINK-15559
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 1.10.0
>Reporter: Benchao Li
>Assignee: Benchao Li
>Priority: Major
>
> For example, in 
>  * dev/table/functions/systemFunctions.md
>  * dev/table/functions/systemFunctions.zh.md
>  * dev/table/sourceSinks.md
>  * dev/table/sourceSinks.zh.md
>  * dev/table/sql/queries.zh.md
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] Type of keys should be `BinaryRow`…

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] 
Type of keys should be `BinaryRow`…
URL: https://github.com/apache/flink/pull/10815#issuecomment-572566376
 
 
   
   ## CI report:
   
   * 19a4290f709495491fe460037c8c31d106984ea8 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143732723) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4229)
 
   * c3ef5ea345a343170806de8112163edb7df31f69 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144110200) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4284)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Caizhi Weng updated FLINK-15570:

Description: 
Currently we only support inserting into empty csv tables. It would be nice for 
a new user from traditional database to try out Flink if we support inserting 
into non-empty csv tables.

In other words, it would be nice if the following SQL produces a valid result:
{code:java}
CREATE TABLE myTable(
a INT,
b DOUBLE
) WITH (
'connector.type' = 'filesystem',
'connector.path' = an existing csv file,
'format.type' = 'csv',
'format.derive-schema' = 'true'
);

SELECT * FROM myTable; // produces 3 rows

INSERT INTO myTable VALUES (4, 4.4);

SELECT * FROM myTable; // produces 4 rows{code}
We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
NO_OVERWRITE mode we can only insert into empty csv tables, while in OVERWRITE 
mode inserting into a csv table will wipe all existing data in it (in the above 
example, the last SELECT will produce only 1 instead of 4 rows) which is really 
strange for a mere INSERT operation.

We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
mode to something like appending to files. Supporting specifying write mode in 
the WITH clause is also needed.

  was:
Currently we only support inserting into empty csv tables. It would be nice for 
a new user from traditional database to try out Flink if we support inserting 
into non-empty csv tables.

In other words, it would be nice if the following SQL produces a valid result:
{code:java}
CREATE TABLE myTable(
a INT,
b DOUBLE
) WITH (
'connector.type' = 'filesystem',
'connector.path' = an existing csv file,
'format.type' = 'csv',
'format.derive-schema' = 'true'
);

SELECT * FROM myTable; // produces 3 rows

INSERT INTO myTable VALUES (4, 4.4);

SELECT * FROM myTable; // produces 4 rows{code}
We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
NO_OVERWRITE mode we can only insert into empty csv tables, while in OVERWRITE 
mode inserting into a csv table will wipe all existing data in it (in the above 
example, the last SELECT will produce only 1 instead of 4 rows) which is really 
strange for a mere INSERT operation. We need to add a new APPEND write mode, or 
change the behavior for OVERWRITE mode to something like appending to files.


> Support inserting into non-empty csv tables
> ---
>
> Key: FLINK-15570
> URL: https://issues.apache.org/jira/browse/FLINK-15570
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
> ORC, SequenceFile)
>Reporter: Caizhi Weng
>Priority: Minor
>
> Currently we only support inserting into empty csv tables. It would be nice 
> for a new user from traditional database to try out Flink if we support 
> inserting into non-empty csv tables.
> In other words, it would be nice if the following SQL produces a valid result:
> {code:java}
> CREATE TABLE myTable(
> a INT,
> b DOUBLE
> ) WITH (
> 'connector.type' = 'filesystem',
> 'connector.path' = an existing csv file,
> 'format.type' = 'csv',
> 'format.derive-schema' = 'true'
> );
> SELECT * FROM myTable; // produces 3 rows
> INSERT INTO myTable VALUES (4, 4.4);
> SELECT * FROM myTable; // produces 4 rows{code}
> We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
> NO_OVERWRITE mode we can only insert into empty csv tables, while in 
> OVERWRITE mode inserting into a csv table will wipe all existing data in it 
> (in the above example, the last SELECT will produce only 1 instead of 4 rows) 
> which is really strange for a mere INSERT operation.
> We need to add a new APPEND write mode, or change the behavior for OVERWRITE 
> mode to something like appending to files. Supporting specifying write mode 
> in the WITH clause is also needed.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15570) Support inserting into non-empty csv tables

2020-01-13 Thread Caizhi Weng (Jira)
Caizhi Weng created FLINK-15570:
---

 Summary: Support inserting into non-empty csv tables
 Key: FLINK-15570
 URL: https://issues.apache.org/jira/browse/FLINK-15570
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / FileSystem, Formats (JSON, Avro, Parquet, 
ORC, SequenceFile)
Reporter: Caizhi Weng


Currently we only support inserting into empty csv tables. It would be nice for 
a new user from traditional database to try out Flink if we support inserting 
into non-empty csv tables.

In other words, it would be nice if the following SQL produces a valid result:
{code:java}
CREATE TABLE myTable(
a INT,
b DOUBLE
) WITH (
'connector.type' = 'filesystem',
'connector.path' = an existing csv file,
'format.type' = 'csv',
'format.derive-schema' = 'true'
);

SELECT * FROM myTable; // produces 3 rows

INSERT INTO myTable VALUES (4, 4.4);

SELECT * FROM myTable; // produces 4 rows{code}
We currently only have two write modes, namely NO_OVERWRITE and OVERWRITE. In 
NO_OVERWRITE mode we can only insert into empty csv tables, while in OVERWRITE 
mode inserting into a csv table will wipe all existing data in it (in the above 
example, the last SELECT will produce only 1 instead of 4 rows) which is really 
strange for a mere INSERT operation. We need to add a new APPEND write mode, or 
change the behavior for OVERWRITE mode to something like appending to files.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-14837) Support Function DDL in TableEnvironment

2020-01-13 Thread Xianxun Ye (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014141#comment-17014141
 ] 

Xianxun Ye commented on FLINK-14837:


Is there a development plan for this issue?

Will it be resolved in 1.10.0 version?

> Support Function DDL in TableEnvironment
> 
>
> Key: FLINK-14837
> URL: https://issues.apache.org/jira/browse/FLINK-14837
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: Zhenqiu Huang
>Assignee: Zhenqiu Huang
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15569) Incorrect sample code in udf document

2020-01-13 Thread Jeff Zhang (Jira)
Jeff Zhang created FLINK-15569:
--

 Summary: Incorrect sample code in udf document
 Key: FLINK-15569
 URL: https://issues.apache.org/jira/browse/FLINK-15569
 Project: Flink
  Issue Type: Improvement
  Components: Documentation
Affects Versions: 1.10.0
Reporter: Jeff Zhang
 Attachments: image-2020-01-13-16-59-00-022.png

Should use JTuple2 instead of JTuple1

 !image-2020-01-13-16-59-00-022.png! 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on issue #10841: [FLINK-15565][table-planner-blink] Fix error conversion of TinyInt, SmallInt literals between Flink and Calcite

2020-01-13 Thread GitBox
flinkbot commented on issue #10841: [FLINK-15565][table-planner-blink] Fix 
error conversion of TinyInt,SmallInt literals between Flink and Calcite
URL: https://github.com/apache/flink/pull/10841#issuecomment-573564277
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit d3abca3b9c9ca0ad030f7e208843502d14ff13a5 (Mon Jan 13 
08:57:31 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #10840: [FLINK-15444][runtime] Make the component AbstractInvokable in CheckpointBarrierHandler NonNull

2020-01-13 Thread GitBox
flinkbot commented on issue #10840: [FLINK-15444][runtime] Make the component 
AbstractInvokable in CheckpointBarrierHandler NonNull
URL: https://github.com/apache/flink/pull/10840#issuecomment-573563119
 
 
   
   ## CI report:
   
   * 61e9e675667e05e44e9c4b5ffd7565f60aec6d62 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10837: [FLINK-15541] Fix unstable case FlinkKinesisConsumerTest

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10837: [FLINK-15541] Fix unstable case 
FlinkKinesisConsumerTest
URL: https://github.com/apache/flink/pull/10837#issuecomment-573544281
 
 
   
   ## CI report:
   
   * 4d045d632eb72a162e5d56aa52a6819042d2f1aa Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/144107145) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4281)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot commented on issue #10839: [FLINK-15567][table][docs] Add documentation for INSERT statements for Flink SQL

2020-01-13 Thread GitBox
flinkbot commented on issue #10839: [FLINK-15567][table][docs] Add 
documentation for INSERT statements for Flink SQL
URL: https://github.com/apache/flink/pull/10839#issuecomment-573563076
 
 
   
   ## CI report:
   
   * c5595ea0c994dcf664cb68ec45a6f500e705a3b7 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15565:
---
Labels: pull-request-available  (was: )

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|380|65569|4294967425|0.94|28.93|false|sarah robinson|2013-03-01 
> 09:11:58.703176|88.81|xylophone band
>  66|478|65669|4294967339|23.66|38.34|true|yuri carson|2013-03-01 
> 09:11:58.703228|64.68|opthamology
>  1

[GitHub] [flink] JingsongLi opened a new pull request #10841: [FLINK-15565][table-planner-blink] Fix error conversion of TinyInt, SmallInt literals between Flink and Calcite

2020-01-13 Thread GitBox
JingsongLi opened a new pull request #10841: [FLINK-15565][table-planner-blink] 
Fix error conversion of TinyInt,SmallInt literals between Flink and Calcite
URL: https://github.com/apache/flink/pull/10841
 
 
   
   ## What is the purpose of the change
   
   In ExpressionConverter, we convert Flink expression to Calcite RexNode. For 
literal, Flink TinyInt/SmallInt literals are wrongly converted to Calcite 
Int/Int literals. This lead to incompatible types of expression and result type.
   
   ## Brief change log
   
   Fix tinyint/smallint/int literals in 
`ExpressionConverter.visit(ValueLiteralExpression)`.
   
   ## Verifying this change
   
   - `ExpressionConverter`
   - `OverWindowITCase.testOverWindowWithUDAG11G`
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector:no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] Type of keys should be `BinaryRow`…

2020-01-13 Thread GitBox
flinkbot edited a comment on issue #10815: [FLINK-15537][table-planner-blink] 
Type of keys should be `BinaryRow`…
URL: https://github.com/apache/flink/pull/10815#issuecomment-572566376
 
 
   
   ## CI report:
   
   * 19a4290f709495491fe460037c8c31d106984ea8 Travis: 
[FAILURE](https://travis-ci.com/flink-ci/flink/builds/143732723) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4229)
 
   * c3ef5ea345a343170806de8112163edb7df31f69 Travis: 
[PENDING](https://travis-ci.com/flink-ci/flink/builds/144110200) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4284)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014128#comment-17014128
 ] 

Jingsong Lee commented on FLINK-15565:
--

The bug is:

In {{ExpressionConverter}}, we convert Flink expression to Calcite RexNode. For 
literal, Flink TinyInt/SmallInt literals are wrongly converted to Calcite 
Int/Int literals. This lead to incompatible types of expression and result type.

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|3

[jira] [Updated] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-15565:
-
Priority: Critical  (was: Major)

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Critical
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|380|65569|4294967425|0.94|28.93|false|sarah robinson|2013-03-01 
> 09:11:58.703176|88.81|xylophone band
>  66|478|65669|4294967339|23.66|38.34|true|yuri carson|2013-03-01 
> 09:11:58.703228|64.68|opthamology
>  12|322|65771|4294967545|84.87|10.76|false|sarah allen|2013

[GitHub] [flink] flinkbot commented on issue #10840: [FLINK-15444][runtime] Make the component AbstractInvokable in CheckpointBarrierHandler NonNull

2020-01-13 Thread GitBox
flinkbot commented on issue #10840: [FLINK-15444][runtime] Make the component 
AbstractInvokable in CheckpointBarrierHandler NonNull
URL: https://github.com/apache/flink/pull/10840#issuecomment-573561483
 
 
   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 61e9e675667e05e44e9c4b5ffd7565f60aec6d62 (Mon Jan 13 
08:48:18 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15533) Writing DataStream as text file fails due to output path already exists

2020-01-13 Thread Kostas Kloudas (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15533?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014126#comment-17014126
 ] 

Kostas Kloudas commented on FLINK-15533:


Thanks for the feedback [~lirui]! I opened a PR for the issue.

> Writing DataStream as text file fails due to output path already exists
> ---
>
> Key: FLINK-15533
> URL: https://issues.apache.org/jira/browse/FLINK-15533
> Project: Flink
>  Issue Type: Bug
>  Components: Client / Job Submission
>Affects Versions: 1.10.0
>Reporter: Rui Li
>Assignee: Kostas Kloudas
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.10.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The following program reproduces the issue.
> {code}
> Configuration configuration = GlobalConfiguration.loadConfiguration();
> configuration.set(DeploymentOptions.TARGET, RemoteExecutor.NAME);
> StreamExecutionEnvironment streamEnv = new 
> StreamExecutionEnvironment(configuration);
> DataStream dataStream = streamEnv.fromCollection(Arrays.asList(1,2,3));
> dataStream.writeAsText("hdfs://localhost:8020/tmp/output");
> streamEnv.execute();
> {code}
> The job will fail with the follow error, even though the output path doesn't 
> exist before job submission:
> {noformat}
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.fs.FileAlreadyExistsException):
>  /tmp/output already exists as a directory
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15568) RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 versions

2020-01-13 Thread Andrew.D.lin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15568?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew.D.lin updated FLINK-15568:
-
Summary: RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE 
semantics, in 1.8 versions  (was: RestartPipelinedRegionStrategy: not ensure 
the EXACTLY_ONCE semantics)

> RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics, in 1.8 
> versions
> --
>
> Key: FLINK-15568
> URL: https://issues.apache.org/jira/browse/FLINK-15568
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.8.0, 1.8.1, 1.8.3
>Reporter: Andrew.D.lin
>Priority: Minor
> Attachments: image-2020-01-13-16-40-47-888.png
>
>
> In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
> checkpoint and marked TODO.
> Should we support this feature (region restart) in flink 1.8?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15568) RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics

2020-01-13 Thread Andrew.D.lin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15568?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew.D.lin updated FLINK-15568:
-
Description: 
In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
checkpoint and marked TODO.

Should we support this feature (region restart) in flink 1.8?

  was:
!image-2020-01-13-16-40-47-888.png!

 

In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
checkpoint and marked TODO.

Should we support this feature (region restart) in flink 1.8?


> RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics
> -
>
> Key: FLINK-15568
> URL: https://issues.apache.org/jira/browse/FLINK-15568
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.8.0, 1.8.1, 1.8.3
>Reporter: Andrew.D.lin
>Priority: Minor
> Attachments: image-2020-01-13-16-40-47-888.png
>
>
> In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
> checkpoint and marked TODO.
> Should we support this feature (region restart) in flink 1.8?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-15568) RestartPipelinedRegionStrategy: not ensure the EXACTLY_ONCE semantics

2020-01-13 Thread Andrew.D.lin (Jira)
Andrew.D.lin created FLINK-15568:


 Summary: RestartPipelinedRegionStrategy: not ensure the 
EXACTLY_ONCE semantics
 Key: FLINK-15568
 URL: https://issues.apache.org/jira/browse/FLINK-15568
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Checkpointing
Affects Versions: 1.8.3, 1.8.1, 1.8.0
Reporter: Andrew.D.lin
 Attachments: image-2020-01-13-16-40-47-888.png

!image-2020-01-13-16-40-47-888.png!

 

In 1.8* versions, FailoverRegion.java restart method  not restore from latest 
checkpoint and marked TODO.

Should we support this feature (region restart) in flink 1.8?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-15565) Incompatible types of expression and result type thrown in codegen

2020-01-13 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15565?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014124#comment-17014124
 ] 

Jingsong Lee commented on FLINK-15565:
--

[~ykt836] [~jark] can you assign this to me?

> Incompatible types of expression and result type thrown in codegen
> --
>
> Key: FLINK-15565
> URL: https://issues.apache.org/jira/browse/FLINK-15565
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.10.0
>
>
> *The sql is:*
>  CREATE TABLE `over10k` (
>  t tinyint,
>  si smallint,
>  i int,
>  b bigint,
>  f float,
>  d double,
>  bo boolean,
>  s varchar,
>  ts timestamp,
>  deci decimal(4,2),
>  bin varchar
>  ) WITH (
>  
> 'connector.path'='/daily_regression_batch_hive_1.10/test_window_with_specific_behavior/sources/over10k.csv',
>  'format.empty-column-as-null'='true',
>  'format.field-delimiter'='|',
>  'connector.type'='filesystem',
>  'format.derive-schema'='true',
>  'format.type'='csv'
>  );
>  select s, rank() over (partition by s order by si), sum(b) over (partition 
> by s order by si) from over10k limit 100;
> *The data is :*
>  109|277|65620|4294967305|97.25|7.80|true|nick quirinius|2013-03-01 
> 09:11:58.703226|27.72|undecided
>  93|263|65725|4294967341|6.06|4.12|false|calvin king|2013-03-01 
> 09:11:58.703299|32.44|values clariffication
>  108|383|65629|4294967510|39.55|47.67|false|jessica zipper|2013-03-01 
> 09:11:58.703133|74.23|nap time
>  89|463|65537|4294967493|64.82|13.79|true|ethan white|2013-03-01 
> 09:11:58.703243|89.52|nap time
>  88|372|65645|4294967358|34.48|11.18|true|quinn thompson|2013-03-01 
> 09:11:58.703168|84.86|forestry
>  123|432|65626|4294967435|2.39|16.49|true|david white|2013-03-01 
> 09:11:58.703136|61.24|joggying
>  57|486|65551|4294967397|36.11|9.88|true|katie xylophone|2013-03-01 
> 09:11:58.703142|57.10|zync studies
>  59|343|65787|4294967312|66.89|6.54|true|mike laertes|2013-03-01 
> 09:11:58.703209|27.56|xylophone band
>  74|267|65671|4294967409|21.14|14.64|true|priscilla miller|2013-03-01 
> 09:11:58.703197|89.06|undecided
>  25|336|65587|4294967336|71.01|14.90|true|tom ichabod|2013-03-01 
> 09:11:58.703127|74.32|zync studies
>  48|346|65712|4294967315|45.01|16.08|true|zach brown|2013-03-01 
> 09:11:58.703108|21.68|zync studies
>  84|385|65776|4294967452|35.80|32.13|false|xavier zipper|2013-03-01 
> 09:11:58.703311|99.46|education
>  58|389|65766|4294967416|95.55|20.62|false|sarah miller|2013-03-01 
> 09:11:58.703215|70.92|history
>  22|403|65565|4294967381|99.65|35.42|false|yuri johnson|2013-03-01 
> 09:11:58.703154|94.47|geology
>  55|428|65733|4294967535|99.54|5.35|false|jessica king|2013-03-01 
> 09:11:58.703233|30.30|forestry
>  117|410|65706|4294967391|50.15|0.21|false|quinn johnson|2013-03-01 
> 09:11:58.703248|65.99|yard duty
>  95|423|65573|4294967378|47.59|17.37|true|alice robinson|2013-03-01 
> 09:11:58.703133|54.57|linguistics
>  87|332|65748|4294967320|19.83|41.67|false|fred ellison|2013-03-01 
> 09:11:58.703289|79.02|mathematics
>  114|263|65674|4294967405|84.44|33.18|true|victor van buren|2013-03-01 
> 09:11:58.703092|63.74|linguistics
>  5|369|65780|4294967488|92.02|38.59|true|zach polk|2013-03-01 
> 09:11:58.703271|67.29|yard duty
>  -3|430|65667|4294967469|65.50|40.46|true|yuri xylophone|2013-03-01 
> 09:11:58.703258|30.94|american history
>  120|264|65769|4294967486|89.97|41.18|false|xavier hernandez|2013-03-01 
> 09:11:58.703140|66.89|philosophy
>  107|317|65634|4294967488|5.68|18.89|false|priscilla ichabod|2013-03-01 
> 09:11:58.703196|39.42|joggying
>  29|386|65723|4294967328|71.48|6.13|false|ulysses ichabod|2013-03-01 
> 09:11:58.703215|86.65|xylophone band
>  22|434|65768|4294967543|44.25|27.56|false|tom polk|2013-03-01 
> 09:11:58.703306|12.30|kindergarten
>  -1|274|65755|4294967300|22.01|35.52|false|oscar king|2013-03-01 
> 09:11:58.703141|33.35|chemistry
>  6|365|65603|4294967522|18.51|5.60|false|gabriella king|2013-03-01 
> 09:11:58.703104|34.20|geology
>  97|414|65757|4294967325|31.82|22.37|false|rachel nixon|2013-03-01 
> 09:11:58.703127|61.00|nap time
>  72|448|65538|4294967524|80.09|7.73|true|luke brown|2013-03-01 
> 09:11:58.703090|95.81|american history
>  51|280|65589|4294967486|57.46|23.35|false|zach xylophone|2013-03-01 
> 09:11:58.703299|11.54|education
>  12|447|65583|4294967389|0.98|29.79|true|yuri polk|2013-03-01 
> 09:11:58.703305|1.89|wind surfing
>  -1|360|65539|4294967464|4.08|39.51|false|oscar davidson|2013-03-01 
> 09:11:58.703144|59.47|nap time
>  0|380|65569|4294967425|0.94|28.93|false|sarah robinson|2013-03-01 
> 09:11:58.703176|88.81|xylophone band
>  66|478|65669|4294967339|23.66|38.34|true|yuri carson|2013-03-01 
> 09:11:58.703228|64.68|opthamolog

[jira] [Commented] (FLINK-15559) Missing {{site.baseurl}} in some docs

2020-01-13 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17014123#comment-17014123
 ] 

Jark Wu commented on FLINK-15559:
-

Why the links are broken because of changing base_url ? 
AFAIK, if a link is broken, it should always be broken no matter what the 
base_url is.  That says we don't need a baseurl option.

> Missing {{site.baseurl}} in some docs
> -
>
> Key: FLINK-15559
> URL: https://issues.apache.org/jira/browse/FLINK-15559
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 1.10.0
>Reporter: Benchao Li
>Assignee: Benchao Li
>Priority: Major
>
> For example, in 
>  * dev/table/functions/systemFunctions.md
>  * dev/table/functions/systemFunctions.zh.md
>  * dev/table/sourceSinks.md
>  * dev/table/sourceSinks.zh.md
>  * dev/table/sql/queries.zh.md
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zhijiangW opened a new pull request #10840: [FLINK-15444][runtime] Make the component AbstractInvokable in CheckpointBarrierHandler NonNull

2020-01-13 Thread GitBox
zhijiangW opened a new pull request #10840: [FLINK-15444][runtime] Make the 
component AbstractInvokable in CheckpointBarrierHandler NonNull
URL: https://github.com/apache/flink/pull/10840
 
 
   ## What is the purpose of the change
   
   *The current component `AbstractInvokable` in `CheckpointBarrierHandler` is 
annotated as `@Nullable`. Actually in real code path it is passed via the 
constructor and never be null. The nullable annotation is only used for unit 
test purpose. But this way would mislead the real usage in practice and bring 
extra troubles, because you have to alway check whether it is null before usage 
in related processes.*
   
   *We can refactor the related unit tests to implement a dummy 
`AbstractInvokable` for tests and remove the @Nullable annotation from the 
related class constructors.*
   
   
   ## Brief change log
   
 - *Remove `Nullable` annotation from `CheckpointedInputGate`*
 - *Remove `Nullable` annotation from `CheckpointBarrierHandler`*
 - *Introduce `DummyCheckpointInvokable` for tests*
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (yes / **no** / don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15444) Make the component AbstractInvokable in CheckpointBarrierHandler NonNull

2020-01-13 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-15444:
---
Labels: pull-request-available  (was: )

> Make the component AbstractInvokable in CheckpointBarrierHandler NonNull 
> -
>
> Key: FLINK-15444
> URL: https://issues.apache.org/jira/browse/FLINK-15444
> Project: Flink
>  Issue Type: Task
>  Components: Runtime / Checkpointing
>Reporter: zhijiang
>Assignee: zhijiang
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The current component {{AbstractInvokable}} in {{CheckpointBarrierHandler}} 
> is annotated as {{@Nullable}}. Actually in real code path it is passed via 
> the constructor and never be null. The nullable annotation is only used for 
> unit test purpose. But this way would mislead the real usage in practice and 
> bring extra troubles, because you have to alway check whether it is null 
> before usage in related processes.
> We can refactor the related unit tests to implement a dummy 
> {{AbstractInvokable}} for tests and remove the {{@Nullable}} annotation from 
> the related class constructors.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


<    1   2   3   4   5   >