[jira] [Comment Edited] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17119635#comment-17119635
 ] 

Etienne Chauchot edited comment on FLINK-17961 at 6/4/20, 1:34 PM:
---

Thanks Aljoscha for commenting. ES has data streams features but only for time 
series data; the aim of this source is to read all kind of data. Apart from 
data streams, it behaves like a database. You read the content of an index 
(similar to a table) corresponding to the given query (similar to SQL). So, 
regarding streaming changes, if there are changes between 2 read requests, at 
the second the whole index (containing the change) will be read another time. 
Regarding failover: I guess exactly once semantics cannot be guaranteed only at 
least once. Indeed there is no ack mechanism on already read data. Under those 
circumstances, I guess an ES source cannot get into ES. So what should a user 
do to read from ES? Should he send ES requests manually from a Map ?


was (Author: echauchot):
Thanks Aljoscha for commenting. ES has data streams features but only for time 
series data; the aim of this source is to read all kind of data. Apart from 
data streams it behaves like a database. You read the content of an index 
(similar to a table) corresponding to the given query (similar to SQL). So, 
regarding streaming changes, if there are changes between 2 read requests, at 
the second the whole index (containing the change) will be read another time. 
Regarding failover: I guess exactly once semantics cannot be guaranteed only at 
least once. Indeed there is no ack mechanism on already read data. Under those 
circumstances, I guess an ES source cannot get into ES. So what should a user 
do to read from ES? Should he send ES requests manually from a Map ?

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] rkhachatryan commented on a change in pull request #12478: [FLINK-17869][task][checkpointing] Fix race condition when caling ChannelStateWriter.abort

2020-06-04 Thread GitBox


rkhachatryan commented on a change in pull request #12478:
URL: https://github.com/apache/flink/pull/12478#discussion_r435266519



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/SubtaskCheckpointCoordinatorImpl.java
##
@@ -158,6 +159,16 @@ private ChannelStateWriter openChannelStateWriter() {
@Override
public void abortCheckpointOnBarrier(long checkpointId, Throwable 
cause, OperatorChain operatorChain) throws IOException {
LOG.debug("Aborting checkpoint via cancel-barrier {} for task 
{}", checkpointId, taskName);
+   lastCheckpointId = Math.max(lastCheckpointId, checkpointId);
+   Iterator iterator = abortedCheckpointIds.iterator();
+   while (iterator.hasNext()) {
+   long next = iterator.next();
+   if (next < lastCheckpointId) {
+   iterator.remove();
+   } else {
+   break;
+   }
+   }

Review comment:
   This is just an optimization.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rkhachatryan commented on a change in pull request #12478: [FLINK-17869][task][checkpointing] Fix race condition when caling ChannelStateWriter.abort

2020-06-04 Thread GitBox


rkhachatryan commented on a change in pull request #12478:
URL: https://github.com/apache/flink/pull/12478#discussion_r435272756



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/SubtaskCheckpointCoordinatorImpl.java
##
@@ -281,6 +281,8 @@ public void notifyCheckpointAborted(long checkpointId, 
OperatorChain opera
}
}
 
+   channelStateWriter.abort(checkpointId, new 
CancellationException("checkpoint aborted via notification"), false);

Review comment:
   This is just an optimization (possible now with `abort` and 
`cleanup=false`).





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rkhachatryan commented on a change in pull request #12478: [FLINK-17869][task][checkpointing] Fix race condition when caling ChannelStateWriter.abort

2020-06-04 Thread GitBox


rkhachatryan commented on a change in pull request #12478:
URL: https://github.com/apache/flink/pull/12478#discussion_r435273439



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/SubtaskCheckpointCoordinatorImpl.java
##
@@ -281,6 +281,8 @@ public void notifyCheckpointAborted(long checkpointId, 
OperatorChain opera
}
}
 
+   channelStateWriter.abort(checkpointId, new 
CancellationException("checkpoint aborted via notification"), false);

Review comment:
   This is just an optimization (possible now with `abort` and 
`cleanup=false`).





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12352: [FLINK-17717][sql-parser] Throws for DDL create temporary system func…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12352:
URL: https://github.com/apache/flink/pull/12352#issuecomment-634517986


   
   ## CI report:
   
   * 7dc817125d0684987814d45192eab87a41ebc5b5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2716)
 
   * 6d12fc558cc4c24403fc39884d4d7fea41dfffcf Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2735)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12486: [hotfix][docs] Modify the lateness definition according to the descri…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12486:
URL: https://github.com/apache/flink/pull/12486#issuecomment-638848235


   
   ## CI report:
   
   * 21bab4df76688fccc953d68398fc2cad54b0a26a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2738)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-637399261


   
   ## CI report:
   
   * 4d1f156fd758a44f4c394a478b6f37bc2de61c86 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2711)
 
   * 63831f0ba3dc2a6ec31acc7fb1cfd1b51288945b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2721)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12479: [FLINK-17260] Fix StreamingKafkaITCase instabilities

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12479:
URL: https://github.com/apache/flink/pull/12479#issuecomment-638721700


   
   ## CI report:
   
   * 1bef9a6921ec7d86739b7d6821b6f23f8602bd1b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2704)
 
   * 2c7be46e975aaaef574cc3620169a3b3178b310e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12484: [FLINK-18127][tests] Streamline manual execution of Java E2E tests

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12484:
URL: https://github.com/apache/flink/pull/12484#issuecomment-638836075


   
   ## CI report:
   
   * 30b3c3dfdc501e2f8c54877d6a9c950e688dfd09 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2734)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12456: [FLINK-17113][sql-cli] Refactor view support in SQL Client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12456:
URL: https://github.com/apache/flink/pull/12456#issuecomment-638032908


   
   ## CI report:
   
   * 12e3932ed8d74a59c0d4c28441b336284a2dffeb Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2713)
 
   * 028c7804132598c982cdc82506781df5a4fb753a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2736)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12485: [FLINK-18130][hive][fs-connector] File name conflict for different jobs in filesystem/hive sink

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12485:
URL: https://github.com/apache/flink/pull/12485#issuecomment-638848123


   
   ## CI report:
   
   * 7fe7332d4767bbfb996e86f1b59d99c8c4713192 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2737)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125944#comment-17125944
 ] 

Etienne Chauchot commented on FLINK-17961:
--

I just commented in the existing design discussion thread: 
[https://lists.apache.org/thread.html/r33cd907cecfd125ab1164ddc8a4d8e45d6bd3afd332fbb034881b1ff%40%3Cdev.flink.apache.org%3E]

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16713) Support source mode of elasticsearch connector

2020-06-04 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125945#comment-17125945
 ] 

Etienne Chauchot commented on FLINK-16713:
--

I just commented in the existing design discussion thread: 
[https://lists.apache.org/thread.html/r33cd907cecfd125ab1164ddc8a4d8e45d6bd3afd332fbb034881b1ff%40%3Cdev.flink.apache.org%3E]

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18070) Time attribute been materialized after sub graph optimize

2020-06-04 Thread YufeiLiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18070?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125952#comment-17125952
 ] 

YufeiLiu commented on FLINK-18070:
--

[~libenchao] 
process time have already been materialized at begining of convert.
{code:java}
val converter = new RelTimeIndicatorConverter(rexBuilder)
val convertedRoot = rootRel.accept(converter)
{code}


> Time attribute been materialized after sub graph optimize
> -
>
> Key: FLINK-18070
> URL: https://issues.apache.org/jira/browse/FLINK-18070
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: YufeiLiu
>Assignee: YufeiLiu
>Priority: Major
> Fix For: 1.11.0
>
>
> Hi, I want to use window aggregate after create temporary, and has multiple 
> sinks. But throw exception:
> {code:java}
> java.lang.AssertionError: type mismatch:
> ref:
> TIME ATTRIBUTE(PROCTIME) NOT NULL
> input:
> TIMESTAMP(3) NOT NULL
> {code}
> I look into the optimizer logic, there is comment at 
> {{CommonSubGraphBasedOptimizer}}:
> "1. In general, for multi-sinks users tend to use VIEW which is a natural 
> common sub-graph."
> After sub graph optimize, time attribute from source have been convert to 
> basic TIMESTAMP type according to {{FlinkRelTimeIndicatorProgram}}. But my 
> create view sql is simple query, I think didn't need to materialized time 
> attribute in theory.
> Here is my code:
> {code:java}
> // connector.type COLLECTION is for debug use
> tableEnv.sqlUpdate("CREATE TABLE source (\n" +
>   "`ts` AS PROCTIME(),\n" +
>   "`order_type` INT\n" +
>   ") WITH (\n" +
>   "'connector.type' = 'COLLECTION',\n" +
>   "'format.type' = 'json'\n" +
>   ")\n");
> tableEnv.createTemporaryView("source_view", tableEnv.sqlQuery("SELECT * FROM 
> source"));
> tableEnv.sqlUpdate("CREATE TABLE sink (\n" +
>   "`result` BIGINT\n" +
>   ") WITH (\n" +
>   "'connector.type' = 'COLLECTION',\n" +
>   "'format.type' = 'json'\n" +
>   ")\n");
> tableEnv.sqlUpdate("INSERT INTO sink \n" +
>   "SELECT\n" +
>   "COUNT(1)\n" +
>   "FROM\n" +
>   "`source_view`\n" +
>   "WHERE\n" +
>   " `order_type` = 33\n" +
>   "GROUP BY\n" +
>   "TUMBLE(`ts`, INTERVAL '5' SECOND)\n");
> tableEnv.sqlUpdate("INSERT INTO sink \n" +
>   "SELECT\n" +
>   "COUNT(1)\n" +
>   "FROM\n" +
>   "`source_view`\n" +
>   "WHERE\n" +
>   " `order_type` = 34\n" +
>   "GROUP BY\n" +
>   "TUMBLE(`ts`, INTERVAL '5' SECOND)\n");
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12479: [FLINK-17260] Fix StreamingKafkaITCase instabilities

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12479:
URL: https://github.com/apache/flink/pull/12479#issuecomment-638721700


   
   ## CI report:
   
   * 1bef9a6921ec7d86739b7d6821b6f23f8602bd1b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2704)
 
   * 2c7be46e975aaaef574cc3620169a3b3178b310e Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2739)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17829) Add documentation for the new JDBC connector

2020-06-04 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-17829:
---
Labels: pull-request-available  (was: )

> Add documentation for the new JDBC connector
> 
>
> Key: FLINK-17829
> URL: https://issues.apache.org/jira/browse/FLINK-17829
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC, Documentation
>Reporter: Jark Wu
>Assignee: Leonard Xu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] leonardBang opened a new pull request #12487: [FLINK-17829][docs][jdbc] Add documentation for the new JDBC connector.

2020-06-04 Thread GitBox


leonardBang opened a new pull request #12487:
URL: https://github.com/apache/flink/pull/12487


   ## What is the purpose of the change
   
   *This pull request Add documentation for new JDBC connector.*
   
   
   ## Brief change log
   
 - Add new JDBC connector documentation which include  include 
"Dependencies", "Example", "Options", "Data Type Mapping" sections.
   - Fix typo error in old documentation(temporary -> temporal).
   
   
   ## Verifying this change
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): ( no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12487: [FLINK-17829][docs][jdbc] Add documentation for the new JDBC connector.

2020-06-04 Thread GitBox


flinkbot commented on pull request #12487:
URL: https://github.com/apache/flink/pull/12487#issuecomment-638886784


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 91172715268381f9bdaa7b2284fd322a9608a4ca (Thu Jun 04 
14:31:34 UTC 2020)
   
✅no warnings
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] kl0u closed pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


kl0u closed pull request #12455:
URL: https://github.com/apache/flink/pull/12455


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18132) Add documentation for the new CSV format

2020-06-04 Thread Jark Wu (Jira)
Jark Wu created FLINK-18132:
---

 Summary: Add documentation for the new CSV format
 Key: FLINK-18132
 URL: https://issues.apache.org/jira/browse/FLINK-18132
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Formats (JSON, Avro, Parquet, ORC, 
SequenceFile)
Reporter: Jark Wu
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18131) Add documentation for the new JSON format

2020-06-04 Thread Jark Wu (Jira)
Jark Wu created FLINK-18131:
---

 Summary: Add documentation for the new JSON format
 Key: FLINK-18131
 URL: https://issues.apache.org/jira/browse/FLINK-18131
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Formats (JSON, Avro, Parquet, ORC, 
SequenceFile)
Reporter: Jark Wu
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18133) Add documentation for the new Avro format

2020-06-04 Thread Jark Wu (Jira)
Jark Wu created FLINK-18133:
---

 Summary: Add documentation for the new Avro format
 Key: FLINK-18133
 URL: https://issues.apache.org/jira/browse/FLINK-18133
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Formats (JSON, Avro, Parquet, ORC, 
SequenceFile)
Reporter: Jark Wu
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17935) Logs could not show up when deploying Flink on Yarn via "--executor"

2020-06-04 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas updated FLINK-17935:
---
Fix Version/s: 1.12.0

> Logs could not show up when deploying Flink on Yarn via "--executor"
> 
>
> Key: FLINK-17935
> URL: https://issues.apache.org/jira/browse/FLINK-17935
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Yang Wang
>Assignee: Kostas Kloudas
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0, 1.12.0
>
>
> {code:java}
> ./bin/flink run -d -p 5 -e yarn-per-job 
> examples/streaming/WindowJoin.jar{code}
> When we use the {{-e/--executor}} to specify the deploy target to Yarn 
> per-job, the logs could not show up. The root cause is we do not set the 
> logging files in {{ExecutorCLI}}. We only do it in the 
> {{FlinkYarnSessionCli}}.
> If we use {{-m yarn-cluster}}, everything works well.
>  
> Maybe we should move the {{setLogConfigFileInConfig}} to 
> {{YarnClusterDescriptor}} to avoid this problem. cc [~kkl0u]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-17935) Logs could not show up when deploying Flink on Yarn via "--executor"

2020-06-04 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-17935.
--
Resolution: Fixed

Fixed on master with a674b5e181a2af050aee4c4f23a86a6c61eda94f
and on release-1.11 with d22874ef6322bfb774d60b09c022c3dbbe295d70

> Logs could not show up when deploying Flink on Yarn via "--executor"
> 
>
> Key: FLINK-17935
> URL: https://issues.apache.org/jira/browse/FLINK-17935
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Yang Wang
>Assignee: Kostas Kloudas
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> {code:java}
> ./bin/flink run -d -p 5 -e yarn-per-job 
> examples/streaming/WindowJoin.jar{code}
> When we use the {{-e/--executor}} to specify the deploy target to Yarn 
> per-job, the logs could not show up. The root cause is we do not set the 
> logging files in {{ExecutorCLI}}. We only do it in the 
> {{FlinkYarnSessionCli}}.
> If we use {{-m yarn-cluster}}, everything works well.
>  
> Maybe we should move the {{setLogConfigFileInConfig}} to 
> {{YarnClusterDescriptor}} to avoid this problem. cc [~kkl0u]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18134) Add documentation for the Debezium format

2020-06-04 Thread Jark Wu (Jira)
Jark Wu created FLINK-18134:
---

 Summary: Add documentation for the Debezium format
 Key: FLINK-18134
 URL: https://issues.apache.org/jira/browse/FLINK-18134
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Formats (JSON, Avro, Parquet, ORC, 
SequenceFile)
Reporter: Jark Wu
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18135) Add documentation for the Canal format

2020-06-04 Thread Jark Wu (Jira)
Jark Wu created FLINK-18135:
---

 Summary: Add documentation for the Canal format
 Key: FLINK-18135
 URL: https://issues.apache.org/jira/browse/FLINK-18135
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation, Formats (JSON, Avro, Parquet, ORC, 
SequenceFile)
Reporter: Jark Wu
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15331) [umberlla] Create a table from a changelog

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15331:

Summary: [umberlla] Create a table from a changelog  (was: Create a table 
from a changelog)

> [umberlla] Create a table from a changelog
> --
>
> Key: FLINK-15331
> URL: https://issues.apache.org/jira/browse/FLINK-15331
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Timo Walther
>Assignee: Jark Wu
>Priority: Major
> Fix For: 1.11.0
>
>
> This an umbrella issue for the feature of interpreting a changelog as a 
> table. It can be used to collect initial ideas and use cases until a FLIP is 
> proposed.
> FLIP-105: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-105%3A+Support+to+Interpret+and+Emit+Changelog+in+Flink+SQL



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15331) [umbrella] Create a table from a changelog

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-15331:

Summary: [umbrella] Create a table from a changelog  (was: [umberlla] 
Create a table from a changelog)

> [umbrella] Create a table from a changelog
> --
>
> Key: FLINK-15331
> URL: https://issues.apache.org/jira/browse/FLINK-15331
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Timo Walther
>Assignee: Jark Wu
>Priority: Major
> Fix For: 1.11.0
>
>
> This an umbrella issue for the feature of interpreting a changelog as a 
> table. It can be used to collect initial ideas and use cases until a FLIP is 
> proposed.
> FLIP-105: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-105%3A+Support+to+Interpret+and+Emit+Changelog+in+Flink+SQL



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Reopened] (FLINK-15331) Create a table from a changelog

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15331?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reopened FLINK-15331:
-

> Create a table from a changelog
> ---
>
> Key: FLINK-15331
> URL: https://issues.apache.org/jira/browse/FLINK-15331
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / API
>Reporter: Timo Walther
>Assignee: Jark Wu
>Priority: Major
> Fix For: 1.11.0
>
>
> This an umbrella issue for the feature of interpreting a changelog as a 
> table. It can be used to collect initial ideas and use cases until a FLIP is 
> proposed.
> FLIP-105: 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-105%3A+Support+to+Interpret+and+Emit+Changelog+in+Flink+SQL



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong commented on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


wuchong commented on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-638894806


   Build is passed in my repo: 
https://dev.azure.com/imjark/Flink/_build/results?buildId=136&view=results
   Merging...



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong closed pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


wuchong closed pull request #12436:
URL: https://github.com/apache/flink/pull/12436


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17847) ArrayIndexOutOfBoundsException happens when codegen StreamExec operator

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-17847:

Fix Version/s: (was: 1.12.0)

> ArrayIndexOutOfBoundsException happens when codegen StreamExec operator
> ---
>
> Key: FLINK-17847
> URL: https://issues.apache.org/jira/browse/FLINK-17847
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Leonard Xu
>Assignee: Leonard Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> user case:
> {code:java}
> //source table 
> create table json_table( 
> w_es BIGINT, 
> w_type STRING, 
> w_isDdl BOOLEAN,
>  w_data ARRAY account_pay_fee DOUBLE>>,
>  w_ts TIMESTAMP(3), 
> w_table STRING) WITH (
>   'connector.type' = 'kafka',
>   'connector.version' = '0.10',
>   'connector.topic' = 'json-test2',
>   'connector.properties.zookeeper.connect' = 'localhost:2181',
>   'connector.properties.bootstrap.servers' = 'localhost:9092',
>   'connector.properties.group.id' = 'test-jdbc',
>   'connector.startup-mode' = 'earliest-offset',
>   'format.type' = 'json',
>   'format.derive-schema' = 'true'
> )
> // real data:
> {"w_es":1589870637000,"w_type":"INSERT","w_isDdl":false,"w_data":[{"pay_info":"channelId=82&onlineFee=89.0&outTradeNo=0&payId=0&payType=02&rechargeId=4&totalFee=89.0&tradeStatus=success&userId=32590183789575&sign=00","online_fee":"89.0","sign":"00","account_pay_fee":"0.0"}],"w_ts":"2020-05-20T13:58:37.131Z","w_table":"111"}
> //query
> select w_ts, 'test' as city1_id,  w_data[0].pay_info AS cate3_id,
>  w_data as pay_order_id from json_table
> {code}
> ~exception:~
> {code:java}
> //
> Caused by: java.lang.ArrayIndexOutOfBoundsException: 1427848Caused by: 
> java.lang.ArrayIndexOutOfBoundsException: 1427848 at 
> org.apache.flink.table.runtime.util.SegmentsUtil.getByteMultiSegments(SegmentsUtil.java:598)
>  at 
> org.apache.flink.table.runtime.util.SegmentsUtil.getByte(SegmentsUtil.java:590)
>  at 
> org.apache.flink.table.runtime.util.SegmentsUtil.bitGet(SegmentsUtil.java:534)
>  at 
> org.apache.flink.table.dataformat.BinaryArray.isNullAt(BinaryArray.java:117) 
> at StreamExecCalc$10.processElement(Unknown Source)
> {code}
>  
> Looks like in the codegen StreamExecCalc$10 operator some operation visit a 
> '-1' index which should be wrong, this bug exits both in 1.10 and 1.11
>  
> {code:java}
> public class StreamExecCalc$10 extends 
> org.apache.flink.table.runtime.operators.AbstractProcessStreamOperator
> implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> private final Object[] references;
> private final org.apache.flink.table.dataformat.BinaryString str$3 = 
> org.apache.flink.table.dataformat.BinaryString.fromString("test");
> private transient 
> org.apache.flink.table.runtime.typeutils.BaseArraySerializer typeSerializer$5;
> final org.apache.flink.table.dataformat.BoxedWrapperRow out = new 
> org.apache.flink.table.dataformat.BoxedWrapperRow(4);
> private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> public StreamExecCalc$10(
> Object[] references,
> org.apache.flink.streaming.runtime.tasks.StreamTask task,
> org.apache.flink.streaming.api.graph.StreamConfig config,
> org.apache.flink.streaming.api.operators.Output output) throws 
> Exception {
> this.references = references;
> typeSerializer$5 = 
> (((org.apache.flink.table.runtime.typeutils.BaseArraySerializer) 
> references[0]));
> this.setup(task, config, output);
> }
> @Override
> public void open() throws Exception {
> super.open();
> }
> @Override
> public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> org.apache.flink.table.dataformat.BaseRow in1 = 
> (org.apache.flink.table.dataformat.BaseRow) element.getValue();
> org.apache.flink.table.dataformat.SqlTimestamp field$2;
> boolean isNull$2;
> org.apache.flink.table.dataformat.BaseArray field$4;
> boolean isNull$4;
> org.apache.flink.table.dataformat.BaseArray field$6;
> org.apache.flink.table.dataformat.BinaryString field$8;
> boolean isNull$8;
> org.apache.flink.table.dataformat.BinaryString result$9;
> boolean isNull$9;
> isNull$2 = in1.isNullAt(4);
> field$2 = null;
> if (!isNull$2) {
> field$2 = in1.getTimestamp(4, 3);
> }
> isNull$4 = in1.isNullAt(3);
> field$4 = null;
> if (!isNull$4) {
> f

[GitHub] [flink] dianfu opened a new pull request #12488: [FLINK-18126][python] Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread GitBox


dianfu opened a new pull request #12488:
URL: https://github.com/apache/flink/pull/12488


   
   ## What is the purpose of the change
   
   *The implementation of method `exception` and `set_exception` are not 
correct. The purpose of the Python CompletableFuture is to hole the results 
from the Java CompletableFuture. We should expose the exception inside the Java 
CompletableFuture to users instead of allow Python users set the exception.*
   
   
   ## Brief change log
   
 - *Throw exception for the method set_result and set_exception*
 - *Return the exception reported in the Java CompletableFuture for the 
method exception*
   
   ## Verifying this change
   
   Verified mannully.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-17847) ArrayIndexOutOfBoundsException happens when codegen StreamExec operator

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-17847.
---
Resolution: Fixed

- master (1.12.0): 3ee4c1f0a094d7562f4188b9fc93a2313e872cde
- 1.11.0: 79b64fc42598eb799e7e4da28fc2b4ab6e446506

> ArrayIndexOutOfBoundsException happens when codegen StreamExec operator
> ---
>
> Key: FLINK-17847
> URL: https://issues.apache.org/jira/browse/FLINK-17847
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Leonard Xu
>Assignee: Leonard Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> user case:
> {code:java}
> //source table 
> create table json_table( 
> w_es BIGINT, 
> w_type STRING, 
> w_isDdl BOOLEAN,
>  w_data ARRAY account_pay_fee DOUBLE>>,
>  w_ts TIMESTAMP(3), 
> w_table STRING) WITH (
>   'connector.type' = 'kafka',
>   'connector.version' = '0.10',
>   'connector.topic' = 'json-test2',
>   'connector.properties.zookeeper.connect' = 'localhost:2181',
>   'connector.properties.bootstrap.servers' = 'localhost:9092',
>   'connector.properties.group.id' = 'test-jdbc',
>   'connector.startup-mode' = 'earliest-offset',
>   'format.type' = 'json',
>   'format.derive-schema' = 'true'
> )
> // real data:
> {"w_es":1589870637000,"w_type":"INSERT","w_isDdl":false,"w_data":[{"pay_info":"channelId=82&onlineFee=89.0&outTradeNo=0&payId=0&payType=02&rechargeId=4&totalFee=89.0&tradeStatus=success&userId=32590183789575&sign=00","online_fee":"89.0","sign":"00","account_pay_fee":"0.0"}],"w_ts":"2020-05-20T13:58:37.131Z","w_table":"111"}
> //query
> select w_ts, 'test' as city1_id,  w_data[0].pay_info AS cate3_id,
>  w_data as pay_order_id from json_table
> {code}
> ~exception:~
> {code:java}
> //
> Caused by: java.lang.ArrayIndexOutOfBoundsException: 1427848Caused by: 
> java.lang.ArrayIndexOutOfBoundsException: 1427848 at 
> org.apache.flink.table.runtime.util.SegmentsUtil.getByteMultiSegments(SegmentsUtil.java:598)
>  at 
> org.apache.flink.table.runtime.util.SegmentsUtil.getByte(SegmentsUtil.java:590)
>  at 
> org.apache.flink.table.runtime.util.SegmentsUtil.bitGet(SegmentsUtil.java:534)
>  at 
> org.apache.flink.table.dataformat.BinaryArray.isNullAt(BinaryArray.java:117) 
> at StreamExecCalc$10.processElement(Unknown Source)
> {code}
>  
> Looks like in the codegen StreamExecCalc$10 operator some operation visit a 
> '-1' index which should be wrong, this bug exits both in 1.10 and 1.11
>  
> {code:java}
> public class StreamExecCalc$10 extends 
> org.apache.flink.table.runtime.operators.AbstractProcessStreamOperator
> implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> private final Object[] references;
> private final org.apache.flink.table.dataformat.BinaryString str$3 = 
> org.apache.flink.table.dataformat.BinaryString.fromString("test");
> private transient 
> org.apache.flink.table.runtime.typeutils.BaseArraySerializer typeSerializer$5;
> final org.apache.flink.table.dataformat.BoxedWrapperRow out = new 
> org.apache.flink.table.dataformat.BoxedWrapperRow(4);
> private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> public StreamExecCalc$10(
> Object[] references,
> org.apache.flink.streaming.runtime.tasks.StreamTask task,
> org.apache.flink.streaming.api.graph.StreamConfig config,
> org.apache.flink.streaming.api.operators.Output output) throws 
> Exception {
> this.references = references;
> typeSerializer$5 = 
> (((org.apache.flink.table.runtime.typeutils.BaseArraySerializer) 
> references[0]));
> this.setup(task, config, output);
> }
> @Override
> public void open() throws Exception {
> super.open();
> }
> @Override
> public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> org.apache.flink.table.dataformat.BaseRow in1 = 
> (org.apache.flink.table.dataformat.BaseRow) element.getValue();
> org.apache.flink.table.dataformat.SqlTimestamp field$2;
> boolean isNull$2;
> org.apache.flink.table.dataformat.BaseArray field$4;
> boolean isNull$4;
> org.apache.flink.table.dataformat.BaseArray field$6;
> org.apache.flink.table.dataformat.BinaryString field$8;
> boolean isNull$8;
> org.apache.flink.table.dataformat.BinaryString result$9;
> boolean isNull$9;
> isNull$2 = in1.isNullAt(4);
> field$2 = null;
> if (!isNull$2) {
> field$2 = in1.getTimestamp(4, 3);
> }
> i

[jira] [Updated] (FLINK-18126) Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18126:
---
Labels: pull-request-available  (was: )

> Correct the exception handling of the Python CompletableFuture
> --
>
> Key: FLINK-18126
> URL: https://issues.apache.org/jira/browse/FLINK-18126
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python
>Affects Versions: 1.11.0
>Reporter: Dian Fu
>Assignee: Dian Fu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The implementation of method `exception` and `set_exception` are not correct. 
> The purpose of the Python CompletableFuture is to hole the results from the 
> Java CompletableFuture. We should expose the exception inside the Java 
> CompletableFuture to users instead of allow Python users set the exception.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-10195) RabbitMQ Source With Checkpointing Doesn't Backpressure Correctly

2020-06-04 Thread Austin Cawley-Edwards (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-10195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125362#comment-17125362
 ] 

Austin Cawley-Edwards edited comment on FLINK-10195 at 6/4/20, 2:50 PM:


Yes, that is correct, though I think it is still an improvement on what's here 
and allows the users to tune their job according to their needs, but I don't 
think the performance issue that you cite can be fixed with what Rabbit 
provides.

EDIT: we might be able to handle this by updating prefetch counts dynamically 
if the buffer has space and there are many unacked messages waiting to be acked 
on checkpoint, though I think that might be too much for an initial 
implementation.

 

If the user has high volume queues and needs checkpointing for EXACTLY_ONCE/ 
AT_LEAST_ONCE guarantees, they'll need to tune their buffer length and 
checkpointing interval. This could also be an opt-in/ opt-out change if there 
are cases that need it disabled, and we should definitely update the docs[1].

Our company still actively uses Rabbit – I'm happy to build off your PR and 
test it out in our jobs.

 

 

 

[1]: 
[https://ci.apache.org/projects/flink/flink-docs-stable/dev/connectors/rabbitmq.html#rabbitmq-source]


was (Author: austince):
Yes, that is correct, though I think it is still an improvement on what's here 
and allows the users to tune their job according to their needs, but I don't 
think the performance issue that you cite can be fixed with what Rabbit 
provides. 

EDIT: we might be able to handle this by updating prefetch counts dynamically 
if the buffer has space and there are many unacked messages waiting to be acked 
on checkpoint, though I think that might be too much for an initial 
implementation.

 

If the user has high volume queues and needs checkpointing for EXACTLY_ONCE/ 
AT_LEAST_ONCE guarantees, they'll need to tune their buffer length and 
checkpointing interval. This could also be an opt-in/ opt-out change if there 
are cases that need it disabled, and we should definitely update the docs.

Our company still actively uses Rabbit – I'm happy to build off your PR and 
test it out in our jobs.

 

 

 

[1]: 
[https://ci.apache.org/projects/flink/flink-docs-stable/dev/connectors/rabbitmq.html#rabbitmq-source]

> RabbitMQ Source With Checkpointing Doesn't Backpressure Correctly
> -
>
> Key: FLINK-10195
> URL: https://issues.apache.org/jira/browse/FLINK-10195
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors/ RabbitMQ
>Affects Versions: 1.4.0, 1.5.0, 1.5.1, 1.6.0
>Reporter: Luka Jurukovski
>Assignee: Luka Jurukovski
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The connection between the RabbitMQ server and the client does not 
> appropriately back pressure when auto acking is disabled. This becomes very 
> problematic when a downstream process throttles the data processing to slower 
> then RabbitMQ sends the data to the client.
> The difference in records ends up being stored in the flink's heap space, 
> which grows indefinitely (or technically to "Integer Max" Deliveries). 
> Looking at RabbitMQ's metrics the number of unacked messages looks like 
> steadily rising saw tooth shape.
> Upon further invesitgation it looks like this is due to how the 
> QueueingConsumer works, messages are added to the BlockingQueue faster then 
> they are being removed and processed, resulting in the previously described 
> behavior.
> This may be intended behavior, however this isn't explicitly obvious in the 
> documentation or any of the examples I have seen.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12488: [FLINK-18126][python] Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread GitBox


flinkbot commented on pull request #12488:
URL: https://github.com/apache/flink/pull/12488#issuecomment-638899477


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 4f60d1df216f704acb71f6546d295ed5dc640e1f (Thu Jun 04 
14:50:36 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-14971) Make all the non-IO operations in CheckpointCoordinator single-threaded

2020-06-04 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125989#comment-17125989
 ] 

Stephan Ewen commented on FLINK-14971:
--

In the second step I was referring to "committing asynchronously", because that 
is also a blocking operation (write to ZooKeeper).

However, committing asynchronously is complex because there is a time when the 
scheduler can ask for a checkpoint but it is not clear which the latest one is 
(due do async committing).

Concerning the cleanup problem: This should happen asychronously (not block JM 
and not block committing) but it needs to backpressure new checkpoint creation. 
It sounds to me like the best way would be to take this into account when 
triggering checkpoints, as an additional condition. 
For example, under default settings, a new checkpoint can only be triggered if 
no other periodic checkpoint is in progress, and there is no more than one 
checkpoint pending under cleanup.

> Make all the non-IO operations in CheckpointCoordinator single-threaded
> ---
>
> Key: FLINK-14971
> URL: https://issues.apache.org/jira/browse/FLINK-14971
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing
>Reporter: Biao Liu
>Assignee: Biao Liu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Currently the ACK and declined message handling are executed in IO thread. 
> This is the only rest part that non-IO operations are executed in IO thread. 
> It blocks introducing main thread executor for {{CheckpointCoordinator}}. It 
> would be resolved in this task.
> After resolving the ACK and declined message issue, the main thread executor 
> would be introduced into {{CheckpointCoordinator}} to instead of timer 
> thread. However the timer thread would be kept (maybe for a while 
> temporarily) to schedule periodic triggering, since FLINK-13848 is not 
> accepted yet.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-14971) Make all the non-IO operations in CheckpointCoordinator single-threaded

2020-06-04 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-14971?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125989#comment-17125989
 ] 

Stephan Ewen edited comment on FLINK-14971 at 6/4/20, 3:02 PM:
---

In the second step I was referring to "committing asynchronously", because that 
is also a blocking operation (write to ZooKeeper).

However, committing asynchronously is complex because there is a time when the 
scheduler can ask for a checkpoint but it is not clear which the latest one is 
(due do async committing). I think we can only approach this once we have 
support for async restore in the scheduler.

Concerning the cleanup problem: This should happen asychronously (not block JM 
and not block committing) but it needs to backpressure new checkpoint creation. 
It sounds to me like the best way would be to take this into account when 
triggering checkpoints, as an additional condition. 
For example, under default settings, a new checkpoint can only be triggered if 
no other periodic checkpoint is in progress, and there is no more than one 
checkpoint pending under cleanup.


was (Author: stephanewen):
In the second step I was referring to "committing asynchronously", because that 
is also a blocking operation (write to ZooKeeper).

However, committing asynchronously is complex because there is a time when the 
scheduler can ask for a checkpoint but it is not clear which the latest one is 
(due do async committing).

Concerning the cleanup problem: This should happen asychronously (not block JM 
and not block committing) but it needs to backpressure new checkpoint creation. 
It sounds to me like the best way would be to take this into account when 
triggering checkpoints, as an additional condition. 
For example, under default settings, a new checkpoint can only be triggered if 
no other periodic checkpoint is in progress, and there is no more than one 
checkpoint pending under cleanup.

> Make all the non-IO operations in CheckpointCoordinator single-threaded
> ---
>
> Key: FLINK-14971
> URL: https://issues.apache.org/jira/browse/FLINK-14971
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing
>Reporter: Biao Liu
>Assignee: Biao Liu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Currently the ACK and declined message handling are executed in IO thread. 
> This is the only rest part that non-IO operations are executed in IO thread. 
> It blocks introducing main thread executor for {{CheckpointCoordinator}}. It 
> would be resolved in this task.
> After resolving the ACK and declined message issue, the main thread executor 
> would be introduced into {{CheckpointCoordinator}} to instead of timer 
> thread. However the timer thread would be kept (maybe for a while 
> temporarily) to schedule periodic triggering, since FLINK-13848 is not 
> accepted yet.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12435: [FLINK-18059] [sql-client] Fix create/drop catalog statement can not be executed in sql client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12435:
URL: https://github.com/apache/flink/pull/12435#issuecomment-637372360


   
   ## CI report:
   
   * 84524977ee6597667093943fce7c1730ab934c35 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2732)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12475: [FLINK-18076][table sql / client] Sql client uses wrong class loader when parsing queries

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12475:
URL: https://github.com/apache/flink/pull/12475#issuecomment-638593958


   
   ## CI report:
   
   * 41240cd524255ffc7cf63d2bc2ee9ff38f3abcbd UNKNOWN
   * 49c19b93edb74ef702c97a4742f4936145913f35 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2717)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12487: [FLINK-17829][docs][jdbc] Add documentation for the new JDBC connector.

2020-06-04 Thread GitBox


flinkbot commented on pull request #12487:
URL: https://github.com/apache/flink/pull/12487#issuecomment-638911231


   
   ## CI report:
   
   * 91172715268381f9bdaa7b2284fd322a9608a4ca UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12456: [FLINK-17113][sql-cli] Refactor view support in SQL Client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12456:
URL: https://github.com/apache/flink/pull/12456#issuecomment-638032908


   
   ## CI report:
   
   * 028c7804132598c982cdc82506781df5a4fb753a Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2736)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12485: [FLINK-18130][hive][fs-connector] File name conflict for different jobs in filesystem/hive sink

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12485:
URL: https://github.com/apache/flink/pull/12485#issuecomment-638848123


   
   ## CI report:
   
   * 7fe7332d4767bbfb996e86f1b59d99c8c4713192 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2737)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12488: [FLINK-18126][python] Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread GitBox


flinkbot commented on pull request #12488:
URL: https://github.com/apache/flink/pull/12488#issuecomment-638911342


   
   ## CI report:
   
   * 4f60d1df216f704acb71f6546d295ed5dc640e1f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] lsyldliu commented on a change in pull request #12303: [FLINK-17625] [table] Fix ArrayIndexOutOfBoundsException in AppendOnlyTopNFunction

2020-06-04 Thread GitBox


lsyldliu commented on a change in pull request #12303:
URL: https://github.com/apache/flink/pull/12303#discussion_r435333138



##
File path: 
flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/operators/rank/AppendOnlyTopNFunction.java
##
@@ -210,14 +210,19 @@ private void processElementWithoutRowNumber(RowData 
input, Collector ou
RowData lastKey = lastEntry.getKey();
List lastList = (List) 
lastEntry.getValue();
// remove last one
-   RowData lastElement = lastList.remove(lastList.size() - 
1);
-   if (lastList.isEmpty()) {
+   int size = lastList.size();
+   RowData lastElement = null;
+   if (size > 0) {
+   lastElement = lastList.get(size - 1);

Review comment:
   @wuchong  I agree with you, but I think we also should subtract 
buffer.currentTopNum =  buffer.currentTopNum - 1 after remove lastElement, So 
is it appropriate to add a method in TopNBuffer such as `public void 
minusTopNum(int delta) {
currentTopNum -= delta;
}`,
   we should call this method behind `dataState.put(lastKey, new 
ArrayList<>(lastList))`, what do you think it?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-17732) RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17732?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17125998#comment-17125998
 ] 

Robert Metzger commented on FLINK-17732:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2709&view=logs&j=5c8e7682-d68f-54d1-16a2-a09310218a49&t=f508e270-48d6-5f1e-3138-42a17e0714f0

{code}
2020-06-04T11:24:09.0403191Z [INFO] Running 
org.apache.flink.test.checkpointing.UnalignedCheckpointCompatibilityITCase
2020-06-04T11:24:14.5454441Z [ERROR] Tests run: 24, Failures: 0, Errors: 1, 
Skipped: 0, Time elapsed: 289.948 s <<< FAILURE! - in 
org.apache.flink.test.checkpointing.RescalingITCase
2020-06-04T11:24:14.5456157Z [ERROR] 
testSavepointRescalingInBroadcastOperatorState[backend = 
rocksdb](org.apache.flink.test.checkpointing.RescalingITCase)  Time elapsed: 
172.589 s  <<< ERROR!
2020-06-04T11:24:14.5457382Z java.util.concurrent.ExecutionException: 
org.apache.flink.runtime.concurrent.FutureUtils$RetryException: Could not 
complete the operation. Number of retries has been exhausted.
2020-06-04T11:24:14.5458881Zat 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
2020-06-04T11:24:14.5459746Zat 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
2020-06-04T11:24:14.5460639Zat 
org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingPartitionedOperatorState(RescalingITCase.java:470)
2020-06-04T11:24:14.5461672Zat 
org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingInBroadcastOperatorState(RescalingITCase.java:409)
2020-06-04T11:24:14.5462426Zat 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2020-06-04T11:24:14.5463058Zat 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2020-06-04T11:24:14.5463839Zat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2020-06-04T11:24:14.5464724Zat 
java.lang.reflect.Method.invoke(Method.java:498)
2020-06-04T11:24:14.5465464Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2020-06-04T11:24:14.5466319Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2020-06-04T11:24:14.5467115Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2020-06-04T11:24:14.5467884Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2020-06-04T11:24:14.5468854Zat 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
2020-06-04T11:24:14.5469522Zat 
org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
2020-06-04T11:24:14.5470302Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-06-04T11:24:14.5470901Zat 
org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
2020-06-04T11:24:14.5471604Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
2020-06-04T11:24:14.5472335Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
2020-06-04T11:24:14.5473034Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-06-04T11:24:14.5473677Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-06-04T11:24:14.5474377Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-06-04T11:24:14.5475179Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-06-04T11:24:14.5475815Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-06-04T11:24:14.5476440Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-06-04T11:24:14.5476991Zat 
org.junit.runners.Suite.runChild(Suite.java:128)
2020-06-04T11:24:14.5477669Zat 
org.junit.runners.Suite.runChild(Suite.java:27)
2020-06-04T11:24:14.5478277Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-06-04T11:24:14.5478924Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-06-04T11:24:14.5479730Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-06-04T11:24:14.5480375Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-06-04T11:24:14.5481030Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-06-04T11:24:14.5481769Zat 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
2020-06-04T11:24:14.5482511Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T11:24:14.5483177Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-06-04T11:24:14.5483750Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-06-04T11:24:14.5484424Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
2020-06-04T

[jira] [Updated] (FLINK-17732) RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-17732:
---
Affects Version/s: 1.12.0

> RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList
> -
>
> Key: FLINK-17732
> URL: https://issues.apache.org/jira/browse/FLINK-17732
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Priority: Major
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1413&view=logs&j=16ccbdb7-2a3e-53da-36eb-fb718edc424a&t=cf61ce33-6fba-5fbe-2c0c-e41c4013e891
> {code}
> 2020-05-15T12:09:16.9432669Z [ERROR] 
> testSavepointRescalingOutPartitionedOperatorStateList[backend = 
> filesystem](org.apache.flink.test.checkpointing.RescalingITCase)  Time 
> elapsed: 180.189 s  <<< ERROR!
> 2020-05-15T12:09:16.9433577Z java.util.concurrent.TimeoutException
> 2020-05-15T12:09:16.9434238Z  at 
> java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1784)
> 2020-05-15T12:09:16.9435119Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-15T12:09:16.9436062Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingPartitionedOperatorState(RescalingITCase.java:473)
> 2020-05-15T12:09:16.9437313Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList(RescalingITCase.java:427)
> 2020-05-15T12:09:16.9438112Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-15T12:09:16.9438858Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-15T12:09:16.9439611Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-15T12:09:16.9440367Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-15T12:09:16.9441502Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-15T12:09:16.9442020Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-15T12:09:16.9442535Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-15T12:09:16.9442984Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-15T12:09:16.9443557Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-15T12:09:16.9444014Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-15T12:09:16.9444379Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9444901Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-15T12:09:16.9445314Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-15T12:09:16.9445812Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-15T12:09:16.9446289Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9446660Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9447097Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9447478Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9447915Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9448284Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-15T12:09:16.9448689Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-15T12:09:16.9449062Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-15T12:09:16.9449405Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9450063Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9450620Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9451313Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9451741Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9452218Z  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2020-05-15T12:09:16.9452694Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-05-15T12:09:16.9453065Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9453468Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-15T12:09:16.9454027Z  at 
> org.juni

[jira] [Updated] (FLINK-17732) RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-17732:
---
Labels: test-stability  (was: )

> RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList
> -
>
> Key: FLINK-17732
> URL: https://issues.apache.org/jira/browse/FLINK-17732
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Priority: Major
>  Labels: test-stability
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1413&view=logs&j=16ccbdb7-2a3e-53da-36eb-fb718edc424a&t=cf61ce33-6fba-5fbe-2c0c-e41c4013e891
> {code}
> 2020-05-15T12:09:16.9432669Z [ERROR] 
> testSavepointRescalingOutPartitionedOperatorStateList[backend = 
> filesystem](org.apache.flink.test.checkpointing.RescalingITCase)  Time 
> elapsed: 180.189 s  <<< ERROR!
> 2020-05-15T12:09:16.9433577Z java.util.concurrent.TimeoutException
> 2020-05-15T12:09:16.9434238Z  at 
> java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1784)
> 2020-05-15T12:09:16.9435119Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-15T12:09:16.9436062Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingPartitionedOperatorState(RescalingITCase.java:473)
> 2020-05-15T12:09:16.9437313Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList(RescalingITCase.java:427)
> 2020-05-15T12:09:16.9438112Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-15T12:09:16.9438858Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-15T12:09:16.9439611Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-15T12:09:16.9440367Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-15T12:09:16.9441502Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-15T12:09:16.9442020Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-15T12:09:16.9442535Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-15T12:09:16.9442984Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-15T12:09:16.9443557Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-15T12:09:16.9444014Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-15T12:09:16.9444379Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9444901Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-15T12:09:16.9445314Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-15T12:09:16.9445812Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-15T12:09:16.9446289Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9446660Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9447097Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9447478Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9447915Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9448284Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-15T12:09:16.9448689Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-15T12:09:16.9449062Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-15T12:09:16.9449405Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9450063Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9450620Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9451313Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9451741Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9452218Z  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2020-05-15T12:09:16.9452694Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-05-15T12:09:16.9453065Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9453468Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 

[jira] [Updated] (FLINK-17732) RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-17732:
---
Component/s: Tests
 Runtime / Checkpointing

> RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList
> -
>
> Key: FLINK-17732
> URL: https://issues.apache.org/jira/browse/FLINK-17732
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Runtime / Task, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Priority: Major
>  Labels: test-stability
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1413&view=logs&j=16ccbdb7-2a3e-53da-36eb-fb718edc424a&t=cf61ce33-6fba-5fbe-2c0c-e41c4013e891
> {code}
> 2020-05-15T12:09:16.9432669Z [ERROR] 
> testSavepointRescalingOutPartitionedOperatorStateList[backend = 
> filesystem](org.apache.flink.test.checkpointing.RescalingITCase)  Time 
> elapsed: 180.189 s  <<< ERROR!
> 2020-05-15T12:09:16.9433577Z java.util.concurrent.TimeoutException
> 2020-05-15T12:09:16.9434238Z  at 
> java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1784)
> 2020-05-15T12:09:16.9435119Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-15T12:09:16.9436062Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingPartitionedOperatorState(RescalingITCase.java:473)
> 2020-05-15T12:09:16.9437313Z  at 
> org.apache.flink.test.checkpointing.RescalingITCase.testSavepointRescalingOutPartitionedOperatorStateList(RescalingITCase.java:427)
> 2020-05-15T12:09:16.9438112Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-15T12:09:16.9438858Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-15T12:09:16.9439611Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-15T12:09:16.9440367Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-15T12:09:16.9441502Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-15T12:09:16.9442020Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-15T12:09:16.9442535Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-15T12:09:16.9442984Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-15T12:09:16.9443557Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-15T12:09:16.9444014Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-05-15T12:09:16.9444379Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9444901Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-15T12:09:16.9445314Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-15T12:09:16.9445812Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-15T12:09:16.9446289Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9446660Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9447097Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9447478Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9447915Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9448284Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-15T12:09:16.9448689Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-05-15T12:09:16.9449062Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-05-15T12:09:16.9449405Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-15T12:09:16.9450063Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-15T12:09:16.9450620Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-15T12:09:16.9451313Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-15T12:09:16.9451741Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-15T12:09:16.9452218Z  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2020-05-15T12:09:16.9452694Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-05-15T12:09:16.9453065Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-05-15T12:09:16.9453468Z  at 
> 

[GitHub] [flink] rmetzger merged pull request #12458: [FLINK-17404] Make sure netty 3.10.6 is used in flink-runtime

2020-06-04 Thread GitBox


rmetzger merged pull request #12458:
URL: https://github.com/apache/flink/pull/12458


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Resolved] (FLINK-17404) Running HA per-job cluster (rocks, non-incremental) gets stuck killing a non-existing pid in Hadoop 3 build profile

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger resolved FLINK-17404.

Resolution: Fixed

Merged to master (1.12.0) in 
https://github.com/apache/flink/commit/9211cb5eaa98f1c51e1569f92c46f65240b0ecc2
Merged to release-1.11 in 
https://github.com/apache/flink/commit/83644269acdc577d26ed376acbe8068af56edbee

> Running HA per-job cluster (rocks, non-incremental) gets stuck killing a 
> non-existing pid in Hadoop 3 build profile
> ---
>
> Key: FLINK-17404
> URL: https://issues.apache.org/jira/browse/FLINK-17404
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Test Infrastructure, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0
>
> Attachments: 255
>
>
> CI log: https://api.travis-ci.org/v3/job/678609505/log.txt
> {code}
> Waiting for text Completed checkpoint [1-9]* for job 
>  to appear 2 of times in logs...
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> Starting standalonejob daemon on host 
> travis-job-e606668f-b674-49c0-8590-e3508e22b99d.
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> Killed TM @ 18864
> kill: usage: kill [-s sigspec | -n signum | -sigspec] pid | jobspec ... or 
> kill -l [sigspec]
> Killed TM @ 
> No output has been received in the last 10m0s, this potentially indicates a 
> stalled build or something wrong with the build itself.
> Check the details on how to adjust your build configuration on: 
> https://docs.travis-ci.com/user/common-build-problems/#build-times-out-because-no-output-was-received
> The build has been terminated
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] rmetzger commented on pull request #12458: [FLINK-17404] Make sure netty 3.10.6 is used in flink-runtime

2020-06-04 Thread GitBox


rmetzger commented on pull request #12458:
URL: https://github.com/apache/flink/pull/12458#issuecomment-638918828


   Thanks a lot for the review!



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-17404) Running HA per-job cluster (rocks, non-incremental) gets stuck killing a non-existing pid in Hadoop 3 build profile

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger closed FLINK-17404.
--

> Running HA per-job cluster (rocks, non-incremental) gets stuck killing a 
> non-existing pid in Hadoop 3 build profile
> ---
>
> Key: FLINK-17404
> URL: https://issues.apache.org/jira/browse/FLINK-17404
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Test Infrastructure, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0
>
> Attachments: 255
>
>
> CI log: https://api.travis-ci.org/v3/job/678609505/log.txt
> {code}
> Waiting for text Completed checkpoint [1-9]* for job 
>  to appear 2 of times in logs...
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> Starting standalonejob daemon on host 
> travis-job-e606668f-b674-49c0-8590-e3508e22b99d.
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> grep: 
> /home/travis/build/apache/flink/flink-dist/target/flink-1.11-SNAPSHOT-bin/flink-1.11-SNAPSHOT/log/*standalonejob-2*.log:
>  No such file or directory
> Killed TM @ 18864
> kill: usage: kill [-s sigspec | -n signum | -sigspec] pid | jobspec ... or 
> kill -l [sigspec]
> Killed TM @ 
> No output has been received in the last 10m0s, this potentially indicates a 
> stalled build or something wrong with the build itself.
> Check the details on how to adjust your build configuration on: 
> https://docs.travis-ci.com/user/common-build-problems/#build-times-out-because-no-output-was-received
> The build has been terminated
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong commented on a change in pull request #12303: [FLINK-17625] [table] Fix ArrayIndexOutOfBoundsException in AppendOnlyTopNFunction

2020-06-04 Thread GitBox


wuchong commented on a change in pull request #12303:
URL: https://github.com/apache/flink/pull/12303#discussion_r435340328



##
File path: 
flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/operators/rank/AppendOnlyTopNFunction.java
##
@@ -210,14 +210,19 @@ private void processElementWithoutRowNumber(RowData 
input, Collector ou
RowData lastKey = lastEntry.getKey();
List lastList = (List) 
lastEntry.getValue();
// remove last one
-   RowData lastElement = lastList.remove(lastList.size() - 
1);
-   if (lastList.isEmpty()) {
+   int size = lastList.size();
+   RowData lastElement = null;
+   if (size > 0) {
+   lastElement = lastList.get(size - 1);

Review comment:
   Good point. I just find that maybe we can use `TopNBuffer#removeLast` 
directly here. 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger merged pull request #12467: [FLINK-16572] Clean up PubSub connector e2e test

2020-06-04 Thread GitBox


rmetzger merged pull request #12467:
URL: https://github.com/apache/flink/pull/12467


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on pull request #12467: [FLINK-16572] Clean up PubSub connector e2e test

2020-06-04 Thread GitBox


rmetzger commented on pull request #12467:
URL: https://github.com/apache/flink/pull/12467#issuecomment-638919331


   Thank for your the review.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-16572) CheckPubSubEmulatorTest is flaky on Azure

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17126003#comment-17126003
 ] 

Robert Metzger commented on FLINK-16572:


Added more debugging in 
https://github.com/apache/flink/commit/52861e30b65d6b50f24028e31a972d68a9a9f9f8

> CheckPubSubEmulatorTest is flaky on Azure
> -
>
> Key: FLINK-16572
> URL: https://issues.apache.org/jira/browse/FLINK-16572
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub, Tests
>Affects Versions: 1.11.0
>Reporter: Aljoscha Krettek
>Assignee: Richard Deurwaarder
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Log: 
> https://dev.azure.com/aljoschakrettek/Flink/_build/results?buildId=56&view=logs&j=1f3ed471-1849-5d3c-a34c-19792af4ad16&t=ce095137-3e3b-5f73-4b79-c42d3d5f8283&l=7842



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] walterddr commented on pull request #12462: [FLINK-18045] Fix Kerberos credentials checking

2020-06-04 Thread GitBox


walterddr commented on pull request #12462:
URL: https://github.com/apache/flink/pull/12462#issuecomment-638923431


   thanks for the quick update. in general the patch looks good to me. I still 
have only one concern regarding the alternation of 2 public APIs. but IMO 
`HadoopUtils` are more of an internal runtime util that was shared between 
different deployment environments. thus I am not sure whether the standard 
public API alternation rules applies here. 
   
   I would wait for @tillrohrmann for a final review. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12487: [FLINK-17829][docs][jdbc] Add documentation for the new JDBC connector.

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12487:
URL: https://github.com/apache/flink/pull/12487#issuecomment-638911231


   
   ## CI report:
   
   * 91172715268381f9bdaa7b2284fd322a9608a4ca Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2744)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12488: [FLINK-18126][python] Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12488:
URL: https://github.com/apache/flink/pull/12488#issuecomment-638911342


   
   ## CI report:
   
   * 4f60d1df216f704acb71f6546d295ed5dc640e1f Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2745)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #12278: [FLINK-17019][runtime] Fulfill slot requests in request order

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12278:
URL: https://github.com/apache/flink/pull/12278#discussion_r435362084



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/SlotPoolImpl.java
##
@@ -648,26 +648,8 @@ boolean offerSlot(
slotOffer.getResourceProfile(),
taskManagerGateway);
 
-   // check whether we have request waiting for this slot
-   PendingRequest pendingRequest = 
pendingRequests.removeKeyB(allocationID);

Review comment:
   There is this `UnfulfillableSlotRequestException` which is still a fail 
fast route if RM finds that a certain request profile cannot be fulfilled at 
all with any existing slot and cannot be allocated. It is relevant for both 
batch and streaming and bulk as I see. I do not know the whole background of 
this. At first glance, it seems to be a complication but it might be necessary.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #12278: [FLINK-17019][runtime] Fulfill slot requests in request order

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12278:
URL: https://github.com/apache/flink/pull/12278#discussion_r435362084



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/SlotPoolImpl.java
##
@@ -648,26 +648,8 @@ boolean offerSlot(
slotOffer.getResourceProfile(),
taskManagerGateway);
 
-   // check whether we have request waiting for this slot
-   PendingRequest pendingRequest = 
pendingRequests.removeKeyB(allocationID);

Review comment:
   There is this `UnfulfillableSlotRequestException` which is still a fail 
fast route if RM finds that a certain request profile cannot be fulfilled at 
all with any existing slot and cannot be allocated. It is relevant for both 
batch and streaming and bulk as I see. I do not know the whole background of 
this. At first glance, it seems to be a complication but it is probably 
necessary to avoid timeout waiting to cancel everything if it is already clear 
that allocation cannot succeed.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12326: [FLINK-16577] [table-planner-blink] Fix numeric type mismatch error in  column interval relmetadata

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12326:
URL: https://github.com/apache/flink/pull/12326#issuecomment-633786013


   
   ## CI report:
   
   * 0b14ce7cf01c6664a3a98c35dd1c494a7ad4fc32 UNKNOWN
   * 8c198bc1bb0b6be3321a01898e09d7492169b47e Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2725)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #12278: [FLINK-17019][runtime] Fulfill slot requests in request order

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12278:
URL: https://github.com/apache/flink/pull/12278#discussion_r435362084



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/SlotPoolImpl.java
##
@@ -648,26 +648,8 @@ boolean offerSlot(
slotOffer.getResourceProfile(),
taskManagerGateway);
 
-   // check whether we have request waiting for this slot
-   PendingRequest pendingRequest = 
pendingRequests.removeKeyB(allocationID);

Review comment:
   There is this `UnfulfillableSlotRequestException` which is still a fail 
fast route if RM finds that a certain request profile cannot be fulfilled at 
all with any existing slot and cannot be allocated. It is relevant for both 
batch and streaming and bulk as I see. I do not know the whole background of 
this. At first glance, it seems to be a complication but it is probably 
necessary to avoid timeout waiting to cancel everything if it is already clear 
that allocation can never succeed.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12488: [FLINK-18126][python] Correct the exception handling of the Python CompletableFuture

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12488:
URL: https://github.com/apache/flink/pull/12488#issuecomment-638911342


   
   ## CI report:
   
   * 4f60d1df216f704acb71f6546d295ed5dc640e1f Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2745)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #12375: [FLINK-17017][runtime] Implements bulk allocation for physical slots

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12375:
URL: https://github.com/apache/flink/pull/12375#discussion_r435366834



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/BulkSlotProviderImpl.java
##
@@ -0,0 +1,271 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.jobmaster.slotpool;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.time.Time;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.clusterframework.types.ResourceProfile;
+import org.apache.flink.runtime.clusterframework.types.SlotProfile;
+import org.apache.flink.runtime.concurrent.ComponentMainThreadExecutor;
+import org.apache.flink.runtime.concurrent.FutureUtils;
+import org.apache.flink.runtime.jobmaster.SlotInfo;
+import org.apache.flink.runtime.jobmaster.SlotRequestId;
+import org.apache.flink.util.clock.Clock;
+import org.apache.flink.util.clock.SystemClock;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Optional;
+import java.util.Set;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.TimeoutException;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * Default implementation of {@link BulkSlotProvider}.
+ */
+class BulkSlotProviderImpl implements BulkSlotProvider {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(BulkSlotProviderImpl.class);
+
+   private ComponentMainThreadExecutor componentMainThreadExecutor;
+
+   private final SlotSelectionStrategy slotSelectionStrategy;
+
+   private final SlotPool slotPool;
+
+   private final Clock clock;
+
+   private final PhysicalSlotRequestBulkTracker slotRequestBulkTracker;
+
+   BulkSlotProviderImpl(final SlotSelectionStrategy slotSelectionStrategy, 
final SlotPool slotPool) {
+   this(slotSelectionStrategy, slotPool, 
SystemClock.getInstance());
+   }
+
+   @VisibleForTesting
+   BulkSlotProviderImpl(
+   final SlotSelectionStrategy slotSelectionStrategy,
+   final SlotPool slotPool,
+   final Clock clock) {
+
+   this.slotSelectionStrategy = 
checkNotNull(slotSelectionStrategy);
+   this.slotPool = checkNotNull(slotPool);
+   this.clock = checkNotNull(clock);
+
+   this.slotRequestBulkTracker = new 
PhysicalSlotRequestBulkTracker(clock);
+
+   this.componentMainThreadExecutor = new 
ComponentMainThreadExecutor.DummyComponentMainThreadExecutor(
+   "Scheduler is not initialized with proper main thread 
executor. " +
+   "Call to BulkSlotProvider.start(...) 
required.");
+   }
+
+   @Override
+   public void start(final ComponentMainThreadExecutor mainThreadExecutor) 
{
+   this.componentMainThreadExecutor = mainThreadExecutor;
+   }
+
+   @Override
+   public CompletableFuture> 
allocatePhysicalSlots(
+   final Collection 
physicalSlotRequests,
+   final Time timeout) {
+
+   componentMainThreadExecutor.assertRunningInMainThread();
+
+   LOG.debug("Received {} slot requests.", 
physicalSlotRequests.size());
+
+   final PhysicalSlotRequestBulk slotRequestBulk = new 
PhysicalSlotRequestBulk(physicalSlotRequests);
+
+   final List> 
resultFutures = new ArrayList<>(physicalSlotRequests.size());
+   for (PhysicalSlotRequest request : physicalSlotRequests) {
+   final CompletableFuture 
resultFuture =
+   allocatePhysicalSlot(request, 
timeout).thenApply(result -> {
+   slotRequestBulk.markRequestFulfilled(
+   result.getSlotRequestId(),
+   
resul

[GitHub] [flink] azagrebin commented on a change in pull request #12278: [FLINK-17019][runtime] Fulfill slot requests in request order

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12278:
URL: https://github.com/apache/flink/pull/12278#discussion_r435362084



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/SlotPoolImpl.java
##
@@ -648,26 +648,8 @@ boolean offerSlot(
slotOffer.getResourceProfile(),
taskManagerGateway);
 
-   // check whether we have request waiting for this slot
-   PendingRequest pendingRequest = 
pendingRequests.removeKeyB(allocationID);

Review comment:
   There is this `UnfulfillableSlotRequestException` which is still a fail 
fast route if RM finds that a certain request profile cannot be fulfilled at 
all with any existing slot and cannot be allocated. It is relevant for both 
batch and streaming and bulk as I see. I do not know the whole background of 
this. At first glance, this looks to me as an optimisation that complicates 
things a bit at the moment. It is probably necessary to avoid timeout waiting 
to cancel everything if it is already clear that allocation can never succeed.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] leonardBang commented on pull request #12475: [FLINK-18076][table sql / client] Sql client uses wrong class loader when parsing queries

2020-06-04 Thread GitBox


leonardBang commented on pull request #12475:
URL: https://github.com/apache/flink/pull/12475#issuecomment-638946255


   The PR updated @dawidwys , please help take a look.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Assigned] (FLINK-17954) Do not multiplex remote function state into single PersistedTable

2020-06-04 Thread Tzu-Li (Gordon) Tai (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai reassigned FLINK-17954:
---

Assignee: Tzu-Li (Gordon) Tai

> Do not multiplex remote function state into single PersistedTable
> -
>
> Key: FLINK-17954
> URL: https://issues.apache.org/jira/browse/FLINK-17954
> Project: Flink
>  Issue Type: Task
>  Components: Stateful Functions
>Affects Versions: statefun-2.0.1, statefun-2.1.0
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Major
>
> We are currently multiplexing multiple remote function's user value states 
> into a single {{PersistedTable}}, using the state name as the table key.
> This is not nice since:
> - It does not allow individual states to have different properties, such as 
> TTL expiration.
> - We are restricted to only value states for remote functions



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12471: [FLINK-18073][FLINK-18029][avro] Fix AvroRowDataSerializationSchema is not serializable and add IT cases

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12471:
URL: https://github.com/apache/flink/pull/12471#issuecomment-638362836


   
   ## CI report:
   
   * ec7b1367e121a52731e84f0f5b8610b362c4c169 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2730)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12468: [FLINK-18075] Wrap the SerializationSchema in KafkaSerializationSchema in Kafka connector

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12468:
URL: https://github.com/apache/flink/pull/12468#issuecomment-638243685


   
   ## CI report:
   
   * 7bf026a4e23a89df952f6b361ab8fe435c0a612d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2727)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12475: [FLINK-18076][table sql / client] Sql client uses wrong class loader when parsing queries

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12475:
URL: https://github.com/apache/flink/pull/12475#issuecomment-638593958


   
   ## CI report:
   
   * 41240cd524255ffc7cf63d2bc2ee9ff38f3abcbd UNKNOWN
   * 49c19b93edb74ef702c97a4742f4936145913f35 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2717)
 
   * e1ee41cc7063f83dadf380786b60733dec1639ed UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] HuangZhenQiu commented on pull request #11541: [FLINK-15416][network] add task manager netty client retry mechenism

2020-06-04 Thread GitBox


HuangZhenQiu commented on pull request #11541:
URL: https://github.com/apache/flink/pull/11541#issuecomment-638957276


   @pnowojski For your comments.
   1) There is no test that is marked to ignore by this PR. Would you please 
double-check?
   2) Rebased master again.
   3) Add join logic to wait for all of the threads successfully executed.
   
   I will try my best to make the PR needs the requirement so that it will use 
less time to merge it later. As it is needed for our internal release, 
hopefully, it can be merged soon.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18136) Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread Roman Khachatryan (Jira)
Roman Khachatryan created FLINK-18136:
-

 Summary: Don't start channel state writing for savepoints (RPC)
 Key: FLINK-18136
 URL: https://issues.apache.org/jira/browse/FLINK-18136
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Checkpointing, Runtime / Task
Affects Versions: 1.11.0
Reporter: Roman Khachatryan
Assignee: Roman Khachatryan
 Fix For: 1.11.0, 1.12.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zhijiangW commented on pull request #12460: [FLINK-18063][checkpointing] Fix the race condition for aborting current checkpoint in CheckpointBarrierUnaligner

2020-06-04 Thread GitBox


zhijiangW commented on pull request #12460:
URL: https://github.com/apache/flink/pull/12460#issuecomment-638963043


   @flinkbot run azure



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rkhachatryan opened a new pull request #12489: [FLINK-18136] Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread GitBox


rkhachatryan opened a new pull request #12489:
URL: https://github.com/apache/flink/pull/12489


   ## What is the purpose of the change
   
   `StreamTask.triggerCheckpoint` calls `channelStateWriter.start` 
unconditionally.
   For savepoints and when unaligned mode is disabled this is incorrect.
   
   ## Brief change log
   
 - add `subtaskCheckpointCoordinator.initCheckpoint` and use it in 
`StreamTask`
 - replace `ChannelStateWriter` with `SubtaskCheckpointCoordinator` in 
`Barrierhandler` and call `initCheckpoint` there too
   
   ## Verifying this change
   
- Added `SubtaskCheckpointCoordinatorTest.testInitCheckpoint` (unit test)
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? no
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18136) Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18136?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18136:
---
Labels: pull-request-available  (was: )

> Don't start channel state writing for savepoints (RPC)
> --
>
> Key: FLINK-18136
> URL: https://issues.apache.org/jira/browse/FLINK-18136
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Runtime / Task
>Affects Versions: 1.11.0
>Reporter: Roman Khachatryan
>Assignee: Roman Khachatryan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0, 1.12.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12489: [FLINK-18136] Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread GitBox


flinkbot commented on pull request #12489:
URL: https://github.com/apache/flink/pull/12489#issuecomment-638974327


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 5136fc6b80380c9ff9be14c67d6a286dbac118a5 (Thu Jun 04 
16:44:44 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12375: [FLINK-17017][runtime] Implements bulk allocation for physical slots

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12375:
URL: https://github.com/apache/flink/pull/12375#issuecomment-635184560


   
   ## CI report:
   
   * 35ef82ee0bf1d008719a52e8182f9b63f3eddb11 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2729)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12475: [FLINK-18076][table sql / client] Sql client uses wrong class loader when parsing queries

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12475:
URL: https://github.com/apache/flink/pull/12475#issuecomment-638593958


   
   ## CI report:
   
   * 41240cd524255ffc7cf63d2bc2ee9ff38f3abcbd UNKNOWN
   * 49c19b93edb74ef702c97a4742f4936145913f35 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2717)
 
   * e1ee41cc7063f83dadf380786b60733dec1639ed Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2749)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12482: [FLINK-17831][doc] Add documentation for the new Kafka connector

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12482:
URL: https://github.com/apache/flink/pull/12482#issuecomment-638825868


   
   ## CI report:
   
   * c8c39d8552852d4b7fdbd84501052d08d9b0963d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2731)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] pnowojski commented on a change in pull request #12470: [FLINK-18074][checkpoint] Ensure task could fail when exception thrown out on notified of checkpoint completed/aborted

2020-06-04 Thread GitBox


pnowojski commented on a change in pull request #12470:
URL: https://github.com/apache/flink/pull/12470#discussion_r435404303



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/StreamTask.java
##
@@ -919,9 +920,33 @@ public final ExecutorService 
getAsyncOperationsThreadPool() {
 
@Override
public Future notifyCheckpointCompleteAsync(long checkpointId) {
-   return 
mailboxProcessor.getMailboxExecutor(TaskMailbox.MAX_PRIORITY).submit(
-   () -> notifyCheckpointComplete(checkpointId),
-   "checkpoint %d complete", checkpointId);
+   return notifyCheckpointOperation(
+   () -> notifyCheckpointComplete(checkpointId),
+   String.format("checkpoint %d complete", checkpointId));
+   }
+
+   @Override
+   public Future notifyCheckpointAbortAsync(long checkpointId) {
+   return notifyCheckpointOperation(
+   () -> 
subtaskCheckpointCoordinator.notifyCheckpointAborted(checkpointId, 
operatorChain, this::isRunning),
+   String.format("checkpoint %d aborted", checkpointId));
+   }
+
+   private Future notifyCheckpointOperation(RunnableWithException 
runnable, String description) {

Review comment:
   Could you re-use this method in 
`org.apache.flink.streaming.runtime.tasks.StreamTask#triggerCheckpointAsync`?
   
   However I see there is a difference in the used mailbox executor (different 
priority), so if this would cause too much extra complexity I'm fine with 
leaving it as it is.

##
File path: 
flink-streaming-java/src/test/java/org/apache/flink/streaming/runtime/tasks/StreamTaskTest.java
##
@@ -969,6 +970,34 @@ public void testNotifyCheckpointOnClosedOperator() throws 
Throwable {
assertEquals(true, operator.closed.get());
}
 
+   @Test
+   public void testFailToConfirmCheckpointCompleted() throws Exception {
+   testFailToConfirmCheckpointMessage(streamTask -> 
streamTask.notifyCheckpointCompleteAsync(1L));
+   }
+
+   @Test
+   public void testFailToConfirmCheckpointAborted() throws Exception {
+   testFailToConfirmCheckpointMessage(streamTask -> 
streamTask.notifyCheckpointAbortAsync(1L));
+   }
+
+   private void testFailToConfirmCheckpointMessage(Consumer> consumer) throws Exception {
+   FailOnNotifyCheckpointOperator operator = new 
FailOnNotifyCheckpointOperator<>();
+   MultipleInputStreamTaskTestHarnessBuilder builder =
+   new 
MultipleInputStreamTaskTestHarnessBuilder<>(OneInputStreamTask::new, 
BasicTypeInfo.INT_TYPE_INFO)
+   .addInput(BasicTypeInfo.INT_TYPE_INFO);
+   StreamTaskMailboxTestHarness harness = builder
+   .setupOutputForSingletonOperatorChain(operator)
+   .build();
+
+   try {
+   consumer.accept(harness.streamTask);
+   harness.streamTask.runMailboxStep();
+   fail();
+   } catch (ExpectedTestException expected) {
+   // expected exception

Review comment:
   I presume that this test was failing on master branch (without your fix)?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12352: [FLINK-17717][sql-parser] Throws for DDL create temporary system func…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12352:
URL: https://github.com/apache/flink/pull/12352#issuecomment-634517986


   
   ## CI report:
   
   * 6d12fc558cc4c24403fc39884d4d7fea41dfffcf Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2735)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637922636


   
   ## CI report:
   
   * 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2598)
 
   * fef42fb8497d9822a12ab18114ecb0ea4f36509f UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12489: [FLINK-18136] Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread GitBox


flinkbot commented on pull request #12489:
URL: https://github.com/apache/flink/pull/12489#issuecomment-638985663


   
   ## CI report:
   
   * 5136fc6b80380c9ff9be14c67d6a286dbac118a5 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12483: [FLINK-15339][table][docs] Correct the terminology of "Time-windowed Join" to "Interval Join" in Table API & SQL

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12483:
URL: https://github.com/apache/flink/pull/12483#issuecomment-638835966


   
   ## CI report:
   
   * d8400d687e17309f9c69c296fdbf16a1323a07f4 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2733)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18038) StateBackendLoader logs application-defined state before it is fully configured

2020-06-04 Thread Steve Bairos (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18038?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17126082#comment-17126082
 ] 

Steve Bairos commented on FLINK-18038:
--

That's fair. What if we modify the "Configuring application-defined state 
backend with job/cluster config" log line to be something like
{code:java}
log.info("Using job/cluster config to configure application-defined state 
backend: {}", fromApplication); {code}
So that it's clear that this is from before the state backend was configured. 

Then we could still move the "Using application-defined state backend: {}" log 
line to the end of the if block that it's in so that it shows the configured 
version as the version that's actually being used. 

 

 

> StateBackendLoader logs application-defined state before it is fully 
> configured
> ---
>
> Key: FLINK-18038
> URL: https://issues.apache.org/jira/browse/FLINK-18038
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / State Backends
>Affects Versions: 1.9.1
>Reporter: Steve Bairos
>Priority: Trivial
>
> In the 
> [StateBackendLoader|[https://github.com/apache/flink/blob/bb46756b84940a6134910e74406bfaff4f2f37e9/flink-runtime/src/main/java/org/apache/flink/runtime/state/StateBackendLoader.java#L201]],
>  there's this log line:
> {code:java}
> logger.info("Using application-defined state backend: {}", fromApplication); 
> {code}
> It seems like this is inaccurate though because immediately after logging 
> this, if fromApplication is a ConfigurableStateBackend, we call the 
> .configure() function and it is replaced by a newly configured StateBackend. 
> To me, it seems like it would be better if we logged the state backend after 
> it was fully configured. In the current setup, we get confusing logs like 
> this: 
> {code:java}
> 2020-05-29 21:39:44,387 INFO  
> org.apache.flink.streaming.runtime.tasks.StreamTask   - Using 
> application-defined state backend: 
> RocksDBStateBackend{checkpointStreamBackend=File State Backend (checkpoints: 
> 's3://pinterest-montreal/checkpoints/xenon-dev-001-20191210/Xenon/BasicJavaStream',
>  savepoints: 'null', asynchronous: UNDEFINED, fileStateThreshold: -1), 
> localRocksDbDirectories=null, enableIncrementalCheckpointing=UNDEFINED, 
> numberOfTransferingThreads=-1}2020-05-29 21:39:44,387 INFO  
> org.apache.flink.streaming.runtime.tasks.StreamTask   - Configuring 
> application-defined state backend with job/cluster config{code}
> Which makes it ambiguous whether or not settings in our flink-conf.yaml like 
> "state.backend.incremental: true" are being applied properly or not. 
>  
> I can make a diff for the change if there aren't any objections



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] tillrohrmann commented on a change in pull request #12446: [FLINK-16225] Implement user class loading exception handler

2020-06-04 Thread GitBox


tillrohrmann commented on a change in pull request #12446:
URL: https://github.com/apache/flink/pull/12446#discussion_r435420178



##
File path: 
flink-core/src/main/java/org/apache/flink/util/ClassLoaderWithErrorHandler.java
##
@@ -0,0 +1,64 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.util;
+
+import java.net.URL;
+import java.net.URLClassLoader;
+import java.util.function.Consumer;
+
+/**
+ * This class loader accepts a custom handler if an exception occurs in {@link 
#loadClass(String, boolean)}.
+ */
+public abstract class ClassLoaderWithErrorHandler extends URLClassLoader {
+   public static final Consumer EMPTY_EXCEPTION_HANDLER = 
classLoadingException -> {};
+
+   private final Consumer classLoadingExceptionHandler;
+
+   protected ClassLoaderWithErrorHandler(URL[] urls, ClassLoader parent) {
+   this(urls, parent, EMPTY_EXCEPTION_HANDLER);
+   }
+
+   protected ClassLoaderWithErrorHandler(
+   URL[] urls,
+   ClassLoader parent,
+   Consumer classLoadingExceptionHandler) {
+   super(urls, parent);
+   this.classLoadingExceptionHandler = 
classLoadingExceptionHandler;
+   }
+
+   @SuppressWarnings("FinalMethod")

Review comment:
   Why do we need this suppression here? Is it because we make the 
protected method `final`?

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobManagerSharedServices.java
##
@@ -140,7 +141,8 @@ public static JobManagerSharedServices fromConfiguration(
blobServer,

BlobLibraryCacheManager.defaultClassLoaderFactory(

FlinkUserCodeClassLoaders.ResolveOrder.fromString(classLoaderResolveOrder),
-   alwaysParentFirstLoaderPatterns));
+   alwaysParentFirstLoaderPatterns,
+   
ClassLoaderWithErrorHandler.EMPTY_EXCEPTION_HANDLER));

Review comment:
   Shouldn't we also fail if we encounter a metaspace OOM on the 
`JobManager` side?

##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobManagerSharedServices.java
##
@@ -140,7 +141,8 @@ public static JobManagerSharedServices fromConfiguration(
blobServer,

BlobLibraryCacheManager.defaultClassLoaderFactory(

FlinkUserCodeClassLoaders.ResolveOrder.fromString(classLoaderResolveOrder),
-   alwaysParentFirstLoaderPatterns));
+   alwaysParentFirstLoaderPatterns,
+   
ClassLoaderWithErrorHandler.EMPTY_EXCEPTION_HANDLER));

Review comment:
   Why are we using an `EMPTY_EXCEPTION_HANDLER` here?

##
File path: 
flink-core/src/test/java/org/apache/flink/util/ClassLoaderWithErrorHandlerTest.java
##
@@ -0,0 +1,66 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.util;
+
+import org.junit.Test;
+
+import java.io.IOException;
+import java.net.URL;
+import java.net.URLClassLoader;
+import java.util.concurrent.atomic.AtomicReference;
+
+import static org.hamcrest.CoreMatchers.is;
+import static o

[jira] [Updated] (FLINK-18137) JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint fails with AskTimeoutException

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18137?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-18137:
---
Labels: test-stability  (was: )

> JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint fails with 
> AskTimeoutException
> 
>
> Key: FLINK-18137
> URL: https://issues.apache.org/jira/browse/FLINK-18137
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Runtime / Task
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.11.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2747&view=logs&j=5c8e7682-d68f-54d1-16a2-a09310218a49&t=45cc9205-bdb7-5b54-63cd-89fdc0983323
> {code}
> 2020-06-04T16:17:20.4404189Z [ERROR] Tests run: 4, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 14.352 s <<< FAILURE! - in 
> org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase
> 2020-06-04T16:17:20.4405548Z [ERROR] 
> testStopJobAfterSavepoint(org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase)
>   Time elapsed: 10.058 s  <<< ERROR!
> 2020-06-04T16:17:20.4407342Z java.util.concurrent.ExecutionException: 
> java.util.concurrent.TimeoutException: Invocation of public default 
> java.util.concurrent.CompletableFuture 
> org.apache.flink.runtime.webmonitor.RestfulGateway.triggerSavepoint(org.apache.flink.api.common.JobID,java.lang.String,boolean,org.apache.flink.api.common.time.Time)
>  timed out.
> 2020-06-04T16:17:20.4409562Z  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
> 2020-06-04T16:17:20.4410333Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2020-06-04T16:17:20.4411259Z  at 
> org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase.cancelWithSavepoint(JobMasterTriggerSavepointITCase.java:264)
> 2020-06-04T16:17:20.4412292Z  at 
> org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint(JobMasterTriggerSavepointITCase.java:127)
> 2020-06-04T16:17:20.4413163Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-06-04T16:17:20.4413990Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-06-04T16:17:20.4414783Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-06-04T16:17:20.4415936Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-06-04T16:17:20.4416693Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-06-04T16:17:20.4417632Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-06-04T16:17:20.4418637Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-06-04T16:17:20.4419367Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-06-04T16:17:20.4420118Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-06-04T16:17:20.4420742Z  at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-06-04T16:17:20.4421909Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-06-04T16:17:20.4422493Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-06-04T16:17:20.4423247Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-06-04T16:17:20.4424263Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-06-04T16:17:20.4424876Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-06-04T16:17:20.4426346Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-06-04T16:17:20.4427052Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-06-04T16:17:20.4427772Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-06-04T16:17:20.4428562Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-06-04T16:17:20.4429158Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-06-04T16:17:20.4429861Z  at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-06-04T16:17:20.4430448Z  at 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-06-04T16:17:20.4431060Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-06-04T16:17:20.4431678Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> 2020-06-04T16:17:20.4432513Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Pro

[jira] [Created] (FLINK-18137) JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint fails with AskTimeoutException

2020-06-04 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18137:
--

 Summary: JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint 
fails with AskTimeoutException
 Key: FLINK-18137
 URL: https://issues.apache.org/jira/browse/FLINK-18137
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination, Runtime / Task
Affects Versions: 1.11.0
Reporter: Robert Metzger
 Fix For: 1.11.0


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2747&view=logs&j=5c8e7682-d68f-54d1-16a2-a09310218a49&t=45cc9205-bdb7-5b54-63cd-89fdc0983323

{code}
2020-06-04T16:17:20.4404189Z [ERROR] Tests run: 4, Failures: 0, Errors: 1, 
Skipped: 0, Time elapsed: 14.352 s <<< FAILURE! - in 
org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase
2020-06-04T16:17:20.4405548Z [ERROR] 
testStopJobAfterSavepoint(org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase)
  Time elapsed: 10.058 s  <<< ERROR!
2020-06-04T16:17:20.4407342Z java.util.concurrent.ExecutionException: 
java.util.concurrent.TimeoutException: Invocation of public default 
java.util.concurrent.CompletableFuture 
org.apache.flink.runtime.webmonitor.RestfulGateway.triggerSavepoint(org.apache.flink.api.common.JobID,java.lang.String,boolean,org.apache.flink.api.common.time.Time)
 timed out.
2020-06-04T16:17:20.4409562Zat 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
2020-06-04T16:17:20.4410333Zat 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
2020-06-04T16:17:20.4411259Zat 
org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase.cancelWithSavepoint(JobMasterTriggerSavepointITCase.java:264)
2020-06-04T16:17:20.4412292Zat 
org.apache.flink.runtime.jobmaster.JobMasterTriggerSavepointITCase.testStopJobAfterSavepoint(JobMasterTriggerSavepointITCase.java:127)
2020-06-04T16:17:20.4413163Zat 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2020-06-04T16:17:20.4413990Zat 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2020-06-04T16:17:20.4414783Zat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2020-06-04T16:17:20.4415936Zat 
java.lang.reflect.Method.invoke(Method.java:498)
2020-06-04T16:17:20.4416693Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2020-06-04T16:17:20.4417632Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2020-06-04T16:17:20.4418637Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2020-06-04T16:17:20.4419367Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2020-06-04T16:17:20.4420118Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T16:17:20.4420742Zat 
org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
2020-06-04T16:17:20.4421909Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-06-04T16:17:20.4422493Zat 
org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
2020-06-04T16:17:20.4423247Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
2020-06-04T16:17:20.4424263Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
2020-06-04T16:17:20.4424876Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-06-04T16:17:20.4426346Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-06-04T16:17:20.4427052Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-06-04T16:17:20.4427772Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-06-04T16:17:20.4428562Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-06-04T16:17:20.4429158Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T16:17:20.4429861Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T16:17:20.4430448Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-06-04T16:17:20.4431060Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-06-04T16:17:20.4431678Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
2020-06-04T16:17:20.4432513Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
2020-06-04T16:17:20.4433396Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
2020-06-04T16:17:20.4434298Zat 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
2020-06-04T16:17:20.4440904Zat 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
2020-06-

[jira] [Updated] (FLINK-17384) Support reading hbase conf dir from flink-conf.yaml

2020-06-04 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li updated FLINK-17384:
--
Summary: Support reading hbase conf dir from flink-conf.yaml  (was: support 
read hbase conf dir from flink.conf just like hadoop_conf)

> Support reading hbase conf dir from flink-conf.yaml
> ---
>
> Key: FLINK-17384
> URL: https://issues.apache.org/jira/browse/FLINK-17384
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase, Deployment / Scripts
>Affects Versions: 1.10.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> hi all:
> when user interacts with hbase should do 2 things when using sql
>  # export HBASE_CONF_DIR
>  # add hbase libs to flink_lib(because the hbase connnector doesn't have 
> client's( and others) jar)
> i think it needs to optimise it.
> for 1) we should support read hbase conf dir from flink.conf just like 
> hadoop_conf in  config.sh
> for 2) we should support HBASE_CLASSPATH in  config.sh. In case of jar 
> conflicts such as guava , we also should support flink-hbase-shaded just like 
> hadoop does



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] carp84 closed pull request #12144: [FLINK-17384][flink-dist] support read hbase conf dir from flink.conf and change HBaseConfiguration construction.

2020-06-04 Thread GitBox


carp84 closed pull request #12144:
URL: https://github.com/apache/flink/pull/12144


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18138) KeyedComplexChainTest.testMigrationAndRestore fails with AskTimeoutException on CI

2020-06-04 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18138:
--

 Summary: KeyedComplexChainTest.testMigrationAndRestore fails with 
AskTimeoutException on CI
 Key: FLINK-18138
 URL: https://issues.apache.org/jira/browse/FLINK-18138
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination, Tests
Affects Versions: 1.12.0
Reporter: Robert Metzger


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2723&view=logs&j=5c8e7682-d68f-54d1-16a2-a09310218a49&t=f508e270-48d6-5f1e-3138-42a17e0714f0

{code}
2020-06-04T12:46:11.4458612Z [ERROR] Tests run: 7, Failures: 0, Errors: 1, 
Skipped: 0, Time elapsed: 18.942 s <<< FAILURE! - in 
org.apache.flink.test.state.operator.restore.keyed.KeyedComplexChainTest
2020-06-04T12:46:11.4461142Z [ERROR] testMigrationAndRestore[Migrate Savepoint: 
1.7](org.apache.flink.test.state.operator.restore.keyed.KeyedComplexChainTest)  
Time elapsed: 10.486 s  <<< ERROR!
2020-06-04T12:46:11.4464014Z java.util.concurrent.ExecutionException: 
java.util.concurrent.TimeoutException: Invocation of public default 
java.util.concurrent.CompletableFuture 
org.apache.flink.runtime.webmonitor.RestfulGateway.triggerSavepoint(org.apache.flink.api.common.JobID,java.lang.String,boolean,org.apache.flink.api.common.time.Time)
 timed out.
2020-06-04T12:46:11.4466561Zat 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
2020-06-04T12:46:11.4468016Zat 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
2020-06-04T12:46:11.4469645Zat 
org.apache.flink.test.state.operator.restore.AbstractOperatorRestoreTestBase.migrateJob(AbstractOperatorRestoreTestBase.java:145)
2020-06-04T12:46:11.4475048Zat 
org.apache.flink.test.state.operator.restore.AbstractOperatorRestoreTestBase.testMigrationAndRestore(AbstractOperatorRestoreTestBase.java:107)
2020-06-04T12:46:11.4476558Zat 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2020-06-04T12:46:11.4477687Zat 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2020-06-04T12:46:11.4479033Zat 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2020-06-04T12:46:11.4480310Zat 
java.lang.reflect.Method.invoke(Method.java:498)
2020-06-04T12:46:11.4481369Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2020-06-04T12:46:11.4482508Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2020-06-04T12:46:11.4483663Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2020-06-04T12:46:11.4485020Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2020-06-04T12:46:11.4486066Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T12:46:11.4487061Zat 
org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
2020-06-04T12:46:11.4488097Zat 
org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
2020-06-04T12:46:11.4488951Zat 
org.junit.rules.RunRules.evaluate(RunRules.java:20)
2020-06-04T12:46:11.4489934Zat 
org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
2020-06-04T12:46:11.4493504Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
2020-06-04T12:46:11.4495100Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
2020-06-04T12:46:11.4496193Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-06-04T12:46:11.4497136Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-06-04T12:46:11.4498138Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-06-04T12:46:11.4499404Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-06-04T12:46:11.4500352Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-06-04T12:46:11.4501250Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-06-04T12:46:11.4502194Zat 
org.junit.runners.Suite.runChild(Suite.java:128)
2020-06-04T12:46:11.4502980Zat 
org.junit.runners.Suite.runChild(Suite.java:27)
2020-06-04T12:46:11.4503827Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-06-04T12:46:11.4504861Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-06-04T12:46:11.4505812Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-06-04T12:46:11.4506761Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-06-04T12:46:11.4507953Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-06-04T12:46:11.4508872Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-06-04T12:46:11.4509847Zat 
org.apache.maven.surefire

[jira] [Updated] (FLINK-17384) Support reading hbase conf dir from flink-conf.yaml

2020-06-04 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li updated FLINK-17384:
--
Fix Version/s: (was: 1.11.0)
   1.12.0
   Issue Type: Improvement  (was: Bug)

Changing JIRA type to Improvement since it's not a bug, and change fix version 
to 1.12.0 since we've long passed feature freeze of 1.11.0

> Support reading hbase conf dir from flink-conf.yaml
> ---
>
> Key: FLINK-17384
> URL: https://issues.apache.org/jira/browse/FLINK-17384
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / HBase, Deployment / Scripts
>Affects Versions: 1.10.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> hi all:
> when user interacts with hbase should do 2 things when using sql
>  # export HBASE_CONF_DIR
>  # add hbase libs to flink_lib(because the hbase connnector doesn't have 
> client's( and others) jar)
> i think it needs to optimise it.
> for 1) we should support read hbase conf dir from flink.conf just like 
> hadoop_conf in  config.sh
> for 2) we should support HBASE_CLASSPATH in  config.sh. In case of jar 
> conflicts such as guava , we also should support flink-hbase-shaded just like 
> hadoop does



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12489: [FLINK-18136] Don't start channel state writing for savepoints (RPC)

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12489:
URL: https://github.com/apache/flink/pull/12489#issuecomment-638985663


   
   ## CI report:
   
   * 5136fc6b80380c9ff9be14c67d6a286dbac118a5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2751)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637922636


   
   ## CI report:
   
   * 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2598)
 
   * fef42fb8497d9822a12ab18114ecb0ea4f36509f Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2750)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17384) Support reading hbase conf dir from flink-conf.yaml

2020-06-04 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li updated FLINK-17384:
--
Description: 
Currently when using hbase connector with Flink SQL, below manual steps are 
required:
 # export HBASE_CONF_DIR
 # add hbase libs to flink_lib(because the hbase connnector doesn't have 
client's( and others) jar)

And we should improve this.

For 1) we could support read hbase conf dir from flink-conf.yaml just like 
hadoop/yarn does.

For 2) we should support HBASE_CLASSPATH in config.sh. In case of jar conflicts 
such as guava, we also should support flink-hbase-shaded just like hadoop does

In this JIRA we focus on implementing the 1st proposal.

  was:
hi all:

when user interacts with hbase should do 2 things when using sql
 # export HBASE_CONF_DIR
 # add hbase libs to flink_lib(because the hbase connnector doesn't have 
client's( and others) jar)

i think it needs to optimise it.

for 1) we should support read hbase conf dir from flink.conf just like 
hadoop_conf in  config.sh

for 2) we should support HBASE_CLASSPATH in  config.sh. In case of jar 
conflicts such as guava , we also should support flink-hbase-shaded just like 
hadoop does


Editing the JIRA description to better reflect what the related PR resolves. 
Please feel free to open another one for the 2nd proposal (while I think the 
shaded part would need some discussions and reach consensus before working on 
it) [~jackylau].

> Support reading hbase conf dir from flink-conf.yaml
> ---
>
> Key: FLINK-17384
> URL: https://issues.apache.org/jira/browse/FLINK-17384
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / HBase, Deployment / Scripts
>Affects Versions: 1.10.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> Currently when using hbase connector with Flink SQL, below manual steps are 
> required:
>  # export HBASE_CONF_DIR
>  # add hbase libs to flink_lib(because the hbase connnector doesn't have 
> client's( and others) jar)
> And we should improve this.
> For 1) we could support read hbase conf dir from flink-conf.yaml just like 
> hadoop/yarn does.
> For 2) we should support HBASE_CLASSPATH in config.sh. In case of jar 
> conflicts such as guava, we also should support flink-hbase-shaded just like 
> hadoop does
> In this JIRA we focus on implementing the 1st proposal.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-17384) Support reading hbase conf dir from flink-conf.yaml

2020-06-04 Thread Yu Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yu Li closed FLINK-17384.
-
Release Note: After FLINK-17384 hbase connector supports reading hbase 
configuration directory from flink-conf.yaml through the `env.hbase.conf.dir` 
property when the `HBASE_CONF_DIR` environment variable is not set.
  Resolution: Implemented

Merged into master via 0799b5c20a127110e47439668cf8f8db2e4ecbf3

> Support reading hbase conf dir from flink-conf.yaml
> ---
>
> Key: FLINK-17384
> URL: https://issues.apache.org/jira/browse/FLINK-17384
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / HBase, Deployment / Scripts
>Affects Versions: 1.10.0
>Reporter: jackylau
>Assignee: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> Currently when using hbase connector with Flink SQL, below manual steps are 
> required:
>  # export HBASE_CONF_DIR
>  # add hbase libs to flink_lib(because the hbase connnector doesn't have 
> client's( and others) jar)
> And we should improve this.
> For 1) we could support read hbase conf dir from flink-conf.yaml just like 
> hadoop/yarn does.
> For 2) we should support HBASE_CLASSPATH in config.sh. In case of jar 
> conflicts such as guava, we also should support flink-hbase-shaded just like 
> hadoop does
> In this JIRA we focus on implementing the 1st proposal.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17860) Recursively remove channel state directories

2020-06-04 Thread Roman Khachatryan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17126121#comment-17126121
 ] 

Roman Khachatryan commented on FLINK-17860:
---

After fixing FLINK-17988 there is no immediate need to

Recursively remove channel state directories

However, I think it should be considered for the next release of Unaligned 
Checkpoints.

 

> Recursively remove channel state directories
> 
>
> Key: FLINK-17860
> URL: https://issues.apache.org/jira/browse/FLINK-17860
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Checkpointing
>Affects Versions: 1.11.0
>Reporter: Roman Khachatryan
>Assignee: Roman Khachatryan
>Priority: Critical
> Fix For: 1.12.0
>
>
> With a high degree of parallelism, we end up with n*s number of files in each 
> checkpoint (n = parallelism, s = stages). Writing them if fast (from many 
> subtasks), removing them is slow (from JM).
> This can't be mitigated by state.backend.fs.memory-threshold because most 
> states are ten to hundreds Mb.
>  
> Instead of going through them 1 by 1, we could remove the directory 
> recursively.
>  
> The easiest way is to remove channelStateHandle.discard() calls and use 
> isRecursive=true  in 
> FsCompletedCheckpointStorageLocation.disposeStorageLocation.
> Note: with the current isRecursive=false there will be an exception if there 
> are any files left under that folder.
>  
> This can be extended to other state handles in future as well.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


<    1   2   3   4   5   6   7   >