[jira] [Updated] (FLINK-18602) Support specific offset for topic list for kafka connector in table api

2020-07-14 Thread Shengkai Fang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18602?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shengkai Fang updated FLINK-18602:
--
Description: 
During FLINK-18449, we decide to support topic discovery for kafka connector in 
table api. However, we can only use format 
{{'partition:0,offset:42;partition:1,offset:300'}} to specify the offset for 
single topic. I think the better format is 
{{'topic:topic-1,partition:0,offset:42;topic:topic-2,partition:1,offset:300'}} 
in topic discovery situation.
  
  
  
  

 

  was:
During FLINK-18449, we decide to support topic discovery for kafka connector in 
table api. However, we can only use format 
{{'partition:0,offset:42;partition:1,offset:300'}} to specify the offset for 
single topic. I think the better format is {{'topic:topic-1, 
partition:0,offset:42;topic:topic-2,partition:1,offset:300'}} in topic 
discovery situation.
 
 
 
 

 


> Support specific offset for topic list for kafka connector in table api
> ---
>
> Key: FLINK-18602
> URL: https://issues.apache.org/jira/browse/FLINK-18602
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka, Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Shengkai Fang
>Priority: Major
> Fix For: 1.12.0
>
>
> During FLINK-18449, we decide to support topic discovery for kafka connector 
> in table api. However, we can only use format 
> {{'partition:0,offset:42;partition:1,offset:300'}} to specify the offset for 
> single topic. I think the better format is 
> {{'topic:topic-1,partition:0,offset:42;topic:topic-2,partition:1,offset:300'}}
>  in topic discovery situation.
>   
>   
>   
>   
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] xiaolong-sn commented on a change in pull request #12881: [FLINK-18512][kinesis] Introducing RecordPublisher. Refactor ShardConsumer to use PollingRecordPublisher

2020-07-14 Thread GitBox


xiaolong-sn commented on a change in pull request #12881:
URL: https://github.com/apache/flink/pull/12881#discussion_r454251135



##
File path: 
flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/proxy/KinesisProxyMarker.java
##
@@ -0,0 +1,28 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.connectors.kinesis.proxy;
+
+import org.apache.flink.annotation.Internal;
+
+/**
+ * A marker interface for generic Kinesis Proxy.
+ */
+@Internal

Review comment:
   What does this interface represent besides the existed 
KinesisProxyInterface?

##
File path: 
flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/internals/ShardConsumer.java
##
@@ -379,49 +205,18 @@ private void 
deserializeRecordForCollectionAndUpdateState(UserRecord record)
}
 
/**
-* Calls {@link KinesisProxyInterface#getRecords(String, int)}, while 
also handling unexpected
-* AWS {@link ExpiredIteratorException}s to assure that we get results 
and don't just fail on
-* such occasions. The returned shard iterator within the successful 
{@link GetRecordsResult} should
-* be used for the next call to this method.
-*
-* Note: it is important that this method is not called again before 
all the records from the last result have been
-* fully collected with {@link 
ShardConsumer#deserializeRecordForCollectionAndUpdateState(UserRecord)}, 
otherwise
-* {@link ShardConsumer#lastSequenceNum} may refer to a sub-record in 
the middle of an aggregated record, leading to
-* incorrect shard iteration if the iterator had to be refreshed.
+* Filters out aggregated records that have previously been processed.
+* This method is to support restarting from a partially consumed 
aggregated sequence number.
 *
-* @param shardItr shard iterator to use
-* @param maxNumberOfRecords the maximum number of records to fetch for 
this getRecords attempt
-* @return get records result
-* @throws InterruptedException
+* @param record the record to filter
+* @return {@code true} if the record should be retained
 */
-   private GetRecordsResult getRecords(String shardItr, int 
maxNumberOfRecords) throws Exception {
-   GetRecordsResult getRecordsResult = null;
-   while (getRecordsResult == null) {
-   try {
-   getRecordsResult = kinesis.getRecords(shardItr, 
maxNumberOfRecords);
-
-   // Update millis behind latest so it gets 
reported by the millisBehindLatest gauge
-   Long millisBehindLatest = 
getRecordsResult.getMillisBehindLatest();
-   if (millisBehindLatest != null) {
-   
shardMetricsReporter.setMillisBehindLatest(millisBehindLatest);
-   }
-   } catch (ExpiredIteratorException eiEx) {
-   LOG.warn("Encountered an unexpected expired 
iterator {} for shard {};" +
-   " refreshing the iterator ...", 
shardItr, subscribedShard);
-
-   shardItr = getShardIterator(lastSequenceNum);
-
-   // sleep for the fetch interval before the next 
getRecords attempt with the refreshed iterator
-   if (fetchIntervalMillis != 0) {
-   Thread.sleep(fetchIntervalMillis);
-   }
-   }
+   private boolean filterDeaggregatedRecord(final UserRecord record) {
+   if (lastSequenceNum.isAggregated()) {

Review comment:
   What does this function means? 

##
File path: 
flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/internals/publisher/polling/PollingRecordPublisher.java
##
@@ -0,0 +1,162 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under 

[jira] [Closed] (FLINK-18603) SQL fails with java.lang.IllegalStateException: No operators defined in streaming topology

2020-07-14 Thread Alexander Filipchik (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18603?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Filipchik closed FLINK-18603.
---
Resolution: Not A Problem

Issues was in custom format.

> SQL fails with java.lang.IllegalStateException: No operators defined in 
> streaming topology
> --
>
> Key: FLINK-18603
> URL: https://issues.apache.org/jira/browse/FLINK-18603
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Alexander Filipchik
>Priority: Major
>
> Hi, was playing with 1.11 and found that code that worked in 1.10.1 fails in 
> 1.11.0 with : 
> {code:java}
> Exception in thread "main" java.lang.IllegalStateException: No operators 
> defined in streaming topology. Cannot generate StreamGraph.
>   at 
> org.apache.flink.table.planner.utils.ExecutorUtils.generateStreamGraph(ExecutorUtils.java:47)
>   at 
> org.apache.flink.table.planner.delegation.StreamExecutor.createPipeline(StreamExecutor.java:47)
>   at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.execute(TableEnvironmentImpl.java:1197)
>   at com.css.flink.avro.confluent.table.SqlTest.main(SqlTest.java:53)
> {code}
> code example:
> {code:java}
> StreamExecutionEnvironment bsEnv = 
> StreamExecutionEnvironment.getExecutionEnvironment();
> EnvironmentSettings bsSettings =
> 
> EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
> StreamTableEnvironment tEnv = StreamTableEnvironment.create(bsEnv, 
> bsSettings);
> String createTable =
> String.format(
> "create table EnrichedOrders ("
> + "name VARCHAR,"
> + "proctime AS PROCTIME()"
> + ") with ("
> + "  'connector.type' = 'kafka',"
> + "  'connector.version' = 'universal',"
> + "  'connector.property-version' = '1',"
> + "  'connector.topic' = '%s',"
> + "  'connector.properties.bootstrap.servers' = '%s',"
> + "  'connector.properties.group.id' = '%s',"
> + "  'connector.startup-mode' = 'earliest-offset',"
> + "  'update-mode' = 'append',"
> + "  'format.type' = 'confluent-avro',"
> + "  'format.schema-registry' = '%s'"
> + ")",
> "avro",
> "broker",
> "testSqlLocal",
> "registry");
> tEnv.executeSql(createTable);
> tEnv.toAppendStream(
> tEnv.sqlQuery(
> "select name, sum(*) "
> + "from EnrichedOrders "
> + "GROUP BY TUMBLE(proctime, INTERVAL '1' MINUTE), name"),
> Row.class)
> .print();
> tEnv.execute("testSql");
> {code}
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18582) Add title anchor link for file event_driven.zh.md

2020-07-14 Thread Congxian Qiu(klion26) (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18582?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157881#comment-17157881
 ] 

Congxian Qiu(klion26) commented on FLINK-18582:
---

[~temper] thanks for creating this ticket. could you please share the pages 
which ref this page. If currently there is no pages ref to 
{{event_driven.zh.md}} maybe we could do this late. what do you think?

>From my side, we always encourage to add title anchor links, but I'm sure 
>whether will we accept the change only add the title anchor(if currently no 
>page refs to it)

cc [~jark]

> Add title anchor link for file event_driven.zh.md
> -
>
> Key: FLINK-18582
> URL: https://issues.apache.org/jira/browse/FLINK-18582
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation, Quickstarts
>Reporter: Herman, Li
>Priority: Major
>
> Translated files should add anchor link for sub-titles.
> Otherwise, links from other documentations may not able to point to this 
> specified title. Such as file event_driven.zh.md with title: Side Outputs.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12903: [FLINK-18492][python] Extract the implementation logic of Beam in Operations

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12903:
URL: https://github.com/apache/flink/pull/12903#issuecomment-658519879


   
   ## CI report:
   
   * aa4bb627890b6e46ef36f44528a24aa304bb7064 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4521)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12899: [FLINK-16699][k8s] Support accessing secured services via K8s secrets

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12899:
URL: https://github.com/apache/flink/pull/12899#issuecomment-658162074


   
   ## CI report:
   
   * 77f95464c6036283d77d6e81e2bd21ab27f29d54 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4506)
 
   * 2f0285b88d653612277714e12e96d48e2f82fc51 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4522)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18603) SQL fails with java.lang.IllegalStateException: No operators defined in streaming topology

2020-07-14 Thread Alexander Filipchik (Jira)
Alexander Filipchik created FLINK-18603:
---

 Summary: SQL fails with java.lang.IllegalStateException: No 
operators defined in streaming topology
 Key: FLINK-18603
 URL: https://issues.apache.org/jira/browse/FLINK-18603
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Runtime
Affects Versions: 1.11.0
Reporter: Alexander Filipchik


Hi, was playing with 1.11 and found that code that worked in 1.10.1 fails in 
1.11.0 with : 
{code:java}
Exception in thread "main" java.lang.IllegalStateException: No operators 
defined in streaming topology. Cannot generate StreamGraph.
at 
org.apache.flink.table.planner.utils.ExecutorUtils.generateStreamGraph(ExecutorUtils.java:47)
at 
org.apache.flink.table.planner.delegation.StreamExecutor.createPipeline(StreamExecutor.java:47)
at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.execute(TableEnvironmentImpl.java:1197)
at com.css.flink.avro.confluent.table.SqlTest.main(SqlTest.java:53)

{code}
code example:
{code:java}
StreamExecutionEnvironment bsEnv = 
StreamExecutionEnvironment.getExecutionEnvironment();
EnvironmentSettings bsSettings =

EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build();
StreamTableEnvironment tEnv = StreamTableEnvironment.create(bsEnv, bsSettings);

String createTable =
String.format(
"create table EnrichedOrders ("
+ "name VARCHAR,"
+ "proctime AS PROCTIME()"
+ ") with ("
+ "  'connector.type' = 'kafka',"
+ "  'connector.version' = 'universal',"
+ "  'connector.property-version' = '1',"
+ "  'connector.topic' = '%s',"
+ "  'connector.properties.bootstrap.servers' = '%s',"
+ "  'connector.properties.group.id' = '%s',"
+ "  'connector.startup-mode' = 'earliest-offset',"
+ "  'update-mode' = 'append',"
+ "  'format.type' = 'confluent-avro',"
+ "  'format.schema-registry' = '%s'"
+ ")",
"avro",
"broker",
"testSqlLocal",
"registry");

tEnv.executeSql(createTable);

tEnv.toAppendStream(
tEnv.sqlQuery(
"select name, sum(*) "
+ "from EnrichedOrders "
+ "GROUP BY TUMBLE(proctime, INTERVAL '1' MINUTE), name"),
Row.class)
.print();

tEnv.execute("testSql");
{code}
 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-18594) The link is broken in kafka doc

2020-07-14 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18594?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-18594.

Resolution: Fixed

master: bfdb3c8339347ad59ef5debb71fb171ffa87d4c8

> The link is broken in kafka doc
> ---
>
> Key: FLINK-18594
> URL: https://issues.apache.org/jira/browse/FLINK-18594
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Assignee: Leonard Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> https://ci.apache.org/builders/flink-docs-master/builds/1897/steps/Build%20docs/logs/stdio
> {code}
>   Liquid Exception: Could not find document 
> 'dev/stream/state/checkpointing.md' in tag 'link'. Make sure the document 
> exists and the path is correct. in dev/table/connectors/kafka.zh.md
> Could not find document 'dev/stream/state/checkpointing.md' in tag 'link'.
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi merged pull request #12897: [FLINK-18594][doc] The link is broken in kafka doc

2020-07-14 Thread GitBox


JingsongLi merged pull request #12897:
URL: https://github.com/apache/flink/pull/12897


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18539) StreamExecutionEnvironment#addSource(SourceFunction, TypeInformation) doesn't use the user defined type information

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18539:

Issue Type: Bug  (was: Task)

> StreamExecutionEnvironment#addSource(SourceFunction, TypeInformation) doesn't 
> use the user defined type information
> ---
>
> Key: FLINK-18539
> URL: https://issues.apache.org/jira/browse/FLINK-18539
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.11.0
>Reporter: Jark Wu
>Assignee: Jark Wu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.2, 1.12.0, 1.11.1
>
>
> {code:java}
> class MySource implements SourceFunction, ResultTypeQueryable {
>  TypeInformation getProducedType() {
>return TypeExtractor.createTypeInfo(SourceFunction.class, this.getClass(), 
> 0, null, null);
>  } 
> }
> DataStream ds = tEnv.addSource(new MySource(), Types.ROW(Types.STRING))
> {code}
> The returned {{TypeInformation}} of {{MySource}} is {{GenericTypeInfo}}, not 
> the user given {{RowTypeInfo}}.
> It seems that {{StreamExecutionEnvironment#getTypeInfo}} doesn't use the user 
> given {{typeInfo}} in the highest priority. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18593) Hive bundle jar URLs are broken

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18593?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18593:

Issue Type: Improvement  (was: New Feature)

> Hive bundle jar URLs are broken
> ---
>
> Key: FLINK-18593
> URL: https://issues.apache.org/jira/browse/FLINK-18593
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive, Documentation
>Affects Versions: 1.11.0
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.1
>
>
> we should use [https://repo.maven.apache.org/maven2/] instead



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18602) Support specific offset for topic list for kafka connector in table api

2020-07-14 Thread Shengkai Fang (Jira)
Shengkai Fang created FLINK-18602:
-

 Summary: Support specific offset for topic list for kafka 
connector in table api
 Key: FLINK-18602
 URL: https://issues.apache.org/jira/browse/FLINK-18602
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kafka, Table SQL / API
Affects Versions: 1.11.0
Reporter: Shengkai Fang
 Fix For: 1.12.0


During FLINK-18449, we decide to support topic discovery for kafka connector in 
table api. However, we can only use format 
{{'partition:0,offset:42;partition:1,offset:300'}} to specify the offset for 
single topic. I think the better format is {{'topic:topic-1, 
partition:0,offset:42;topic:topic-2,partition:1,offset:300'}} in topic 
discovery situation.
 
 
 
 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12902: [FLINK-18415][python] Support TableResult#collect in the Python Table…

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12902:
URL: https://github.com/apache/flink/pull/12902#issuecomment-658509076


   
   ## CI report:
   
   * 802dfd4a0a2fb3e1b6e9a31fbc850c18f00c5c2b Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4518)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12899: [FLINK-16699][k8s] Support accessing secured services via K8s secrets

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12899:
URL: https://github.com/apache/flink/pull/12899#issuecomment-658162074


   
   ## CI report:
   
   * 77f95464c6036283d77d6e81e2bd21ab27f29d54 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4506)
 
   * 2f0285b88d653612277714e12e96d48e2f82fc51 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12866:
URL: https://github.com/apache/flink/pull/12866#issuecomment-656576219


   
   ## CI report:
   
   * 5eb302884e390b6099859abcd7955e7f1c4e034e Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4517)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17726) Scheduler should take care of tasks directly canceled by TaskManager

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17726?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-17726:

Fix Version/s: (was: 1.11.1)
   1.11.2

> Scheduler should take care of tasks directly canceled by TaskManager
> 
>
> Key: FLINK-17726
> URL: https://issues.apache.org/jira/browse/FLINK-17726
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Runtime / Task
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Zhu Zhu
>Assignee: Nicholas Jiang
>Priority: Critical
> Fix For: 1.12.0, 1.11.2
>
>
> JobManager will not trigger failure handling when receiving CANCELED task 
> update. 
> This is because CANCELED tasks are usually caused by another FAILED task. 
> These CANCELED tasks will be restarted by the failover process triggered  
> FAILED task.
> However, if a task is directly CANCELED by TaskManager due to its own runtime 
> issue, the task will not be recovered by JM and thus the job would hang.
> This is a potential issue and we should avoid it.
> A possible solution is to let JobManager treat tasks transitioning to 
> CANCELED from all states except from CANCELING as failed tasks. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17075) Add task status reconciliation between TM and JM

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-17075:

Fix Version/s: (was: 1.11.1)
   1.11.2

> Add task status reconciliation between TM and JM
> 
>
> Key: FLINK-17075
> URL: https://issues.apache.org/jira/browse/FLINK-17075
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Till Rohrmann
>Assignee: Chesnay Schepler
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.10.2, 1.12.0, 1.11.2
>
>
> In order to harden the TM and JM communication I suggest to let the 
> {{TaskExecutor}} send the task statuses back to the {{JobMaster}} as part of 
> the heartbeat payload (similar to FLINK-11059). This would allow to reconcile 
> the states of both components in case that a status update message was lost 
> as described by a user on the ML.
> https://lists.apache.org/thread.html/ra9ed70866381f0ef0f4779633346722ccab3dc0d6dbacce04080b74e%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18479) can not commit partition when set partition time

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18479:

Fix Version/s: (was: 1.11.1)
   1.11.2

> can not commit partition when set partition time
> 
>
> Key: FLINK-18479
> URL: https://issues.apache.org/jira/browse/FLINK-18479
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Jun Zhang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.2
>
>
> when we write streaming data to filesystem, and select the 'partition time' , 
> we can not commit the partition when write finished.
>  
> {code:java}
> LocalDateTime partTime = extractor.extract(
>   partitionKeys, extractPartitionValues(new Path(partition)));
> if (watermark > toMills(partTime) + commitDelay) {
>needCommit.add(partition);
>iter.remove();
> }
> {code}
> when we set a not UTC zone, and submit the partition, the method 'toMills' 
> will get the UTC mills ,for example ,in UTC/GMT+08:00 ,the watermark will 
> less than the toMills , so we can not commit the partition forever.
> if we use a local time , not utc , it will be ok in UTC zone and other zone.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18549) flink 1.11 can not commit partition automatically

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18549:

Fix Version/s: (was: 1.11.1)
   1.11.2

> flink 1.11 can not commit partition automatically
> -
>
> Key: FLINK-18549
> URL: https://issues.apache.org/jira/browse/FLINK-18549
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Jun Zhang
>Priority: Major
> Fix For: 1.11.2
>
>
> I use the sql of flink 1.11, read from kafka and writing to hdfs, I found 
> that the partition cannot be submitted automatically. This is my complete 
> code。
> My checkpoint interval is 10s. I think it should be normal that there will be 
> _SUCCESS file under the partition of hdfs every 10s, but in fact there is no
>  
> {code:java}
>StreamExecutionEnvironment bsEnv = 
> StreamExecutionEnvironment.getExecutionEnvironment();
>   bsEnv.enableCheckpointing(1);
>   bsEnv.setParallelism(1);
>   StreamTableEnvironment tEnv = StreamTableEnvironment.create(bsEnv);
>   String sqlSource = "CREATE TABLE  source_kafka (\n" +
>  "appName  STRING,\n" +
>  "appVersion STRING,\n" +
>  "uploadTime STRING\n" +
>  ") WITH (\n" +
>  "  'connector.type' = 'kafka',   \n" +
>  "  'connector.version' = '0.10',\n" +
>  "  'connector.topic' = 'test_topic',\n" +
>  "  'connector.properties.zookeeper.connect' = 
> 'localhost:2181',\n" +
>  "  'connector.properties.bootstrap.servers' = 
> 'localhost:9092',\n" +
>  "  'connector.properties.group.id' = 'testGroup',\n" 
> +
>  "  'format.type'='json',\n" +
>  "  'update-mode' = 'append' )";
>   tEnv.executeSql(sqlSource);
>   String sql = "CREATE TABLE fs_table (\n" +
>"appName  STRING,\n" +
>"appVersion STRING,\n" +
>"uploadTime STRING,\n" +
>"  dt STRING," +
>"  h string" +
>")  PARTITIONED BY (dt,h)  WITH (\n" +
>"  'connector'='filesystem',\n" +
>  "  'path'='hdfs://localhost/tmp/',\n" +
>  " 'sink.partition-commit.policy.kind' = 'success-file', 
> " +
>  "  'format'='orc'\n" +
>  ")";
>   tEnv.executeSql(sql);
>   String insertSql = "insert into  fs_table SELECT appName 
> ,appVersion,uploadTime, " +
>  " DATE_FORMAT(LOCALTIMESTAMP, '-MM-dd'), 
> DATE_FORMAT(LOCALTIMESTAMP, 'HH') FROM source_kafka";
>   tEnv.executeSql(insertSql);
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18376) java.lang.ArrayIndexOutOfBoundsException in RetractableTopNFunction

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18376:

Fix Version/s: (was: 1.11.1)
   1.11.2

> java.lang.ArrayIndexOutOfBoundsException in RetractableTopNFunction
> ---
>
> Key: FLINK-18376
> URL: https://issues.apache.org/jira/browse/FLINK-18376
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: LakeShen
>Assignee: LakeShen
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.2
>
>
> java.lang.ArrayIndexOutOfBoundsException: -1
>   at java.util.ArrayList.elementData(ArrayList.java:422)
>   at java.util.ArrayList.get(ArrayList.java:435)
>   at 
> org.apache.flink.table.runtime.operators.rank.RetractableTopNFunction.retractRecordWithoutRowNumber(RetractableTopNFunction.java:392)
>   at 
> org.apache.flink.table.runtime.operators.rank.RetractableTopNFunction.processElement(RetractableTopNFunction.java:160)
>   at 
> org.apache.flink.table.runtime.operators.rank.RetractableTopNFunction.processElement(RetractableTopNFunction.java:54)
>   at 
> org.apache.flink.streaming.api.operators.KeyedProcessOperator.processElement(KeyedProcessOperator.java:85)
>   at 
> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:173)
>   at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:151)
>   at 
> org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:128)
>   at 
> org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:69)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:311)
>   at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:187)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:487)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:470)
>   at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:707)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:532)
>   at java.lang.Thread.run(Thread.java:748)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18206) The timestamp is displayed incorrectly

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18206?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18206:

Fix Version/s: (was: 1.11.1)
   1.11.2

>  The timestamp is displayed incorrectly
> ---
>
> Key: FLINK-18206
> URL: https://issues.apache.org/jira/browse/FLINK-18206
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.1
>Reporter: JasonLee
>Priority: Minor
> Fix For: 1.11.2
>
>
> I am using the latest Flink version. When I run a scrolling window SQL in SQL 
> client, the time stamp of the printed result will not be correct
>  
> The results are as follows
>  
> + jason 49 2 2020-06-09T07:59:40 2020-06-09T07:59:45
>  + jason 50 2 2020-06-09T07:59:45 2020-06-09T07:59:50
>  + jason 50 2 2020-06-09T07:59:50 2020-06-09T07:59:55
>  + jason 50 2 2020-06-09T07:59:55 2020-06-09T08:00
>  + jason 49 2 2020-06-09T08:00 2020-06-09T08:00:05
>  + jason 50 2 2020-06-09T08:00:05 2020-06-09T08:00:10
>  + jason 50 2 2020-06-09T08:00:10 2020-06-09T08:00:15
>  + jason 50 2 2020-06-09T08:00:15 2020-06-09T08:00:20
>  + jason 49 2 2020-06-09T08:00:20 2020-06-09T08:00:25
>  + jason 50 2 2020-06-09T08:00:25 2020-06-09T08:00:30
>  + jason 50 2 2020-06-09T08:00:30 2020-06-09T08:00:35
>  + jason 49 2 2020-06-09T08:00:35 2020-06-09T08:00:40
>  + jason 51 2 2020-06-09T08:00:40 2020-06-09T08:00:45
>  + jason 50 2 2020-06-09T08:00:45 2020-06-09T08:00:50
>  + jason 49 2 2020-06-09T08:00:50 2020-06-09T08:00:55
>  + jason 50 2 2020-06-09T08:00:55 2020-06-09T08:01
>  + jason 50 2 2020-06-09T08:01 2020-06-09T08:01:05
>  + jason 51 2 2020-06-09T08:01:05 2020-06-09T08:01:10
>  + jason 49 2 2020-06-09T08:01:10 2020-06-09T08:01:15
>  + jason 46 2 2020-06-09T08:01:15 2020-06-09T08:01:20
>  + jason 54 2 2020-06-09T08:01:20 2020-06-09T08:01:25
>  + jason 50 2 2020-06-09T08:01:25 2020-06-09T08:01:30
>  + jason 49 2 2020-06-09T08:01:30 2020-06-09T08:01:35
>  + jason 50 2 2020-06-09T08:01:35 2020-06-09T08:01:40
>  + jason 50 2 2020-06-09T08:01:40 2020-06-09T08:01:45
>  + jason 50 2 2020-06-09T08:01:45 2020-06-09T08:01:50
>  + jason 49 2 2020-06-09T08:01:50 2020-06-09T08:01:55
>  + jason 50 2 2020-06-09T08:01:55 2020-06-09T08:02
>  + jason 49 2 2020-06-09T08:02 2020-06-09T08:02:05
>  + jason 51 2 2020-06-09T08:02:05 2020-06-09T08:02:10
>  + jason 49 2 2020-06-09T08:02:10 2020-06-09T08:02:15
>  + jason 50 2 2020-06-09T08:02:15 2020-06-09T08:02:20
>  + jason 50 2 2020-06-09T08:02:20 2020-06-09T08:02:25
>  + jason 50 2 2020-06-09T08:02:25 2020-06-09T08:02:30
>  + jason 50 2 2020-06-09T08:02:30 2020-06-09T08:02:35
>  + jason 50 2 2020-06-09T08:02:35 2020-06-09T08:02:40
>  + jason 49 2 2020-06-09T08:02:40 2020-06-09T08:02:45
>  + jason 50 2 2020-06-09T08:02:45 2020-06-09T08:02:50
>  + jason 50 2 2020-06-09T08:02:50 2020-06-09T08:02:55
>  + jason 50 2 2020-06-09T08:02:55 2020-06-09T08:03
>  + jason 49 2 2020-06-09T08:03 2020-06-09T08:03:05
>  + jason 51 2 2020-06-09T08:03:05 2020-06-09T08:03:10



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18365) The same sql in a batch env and a streaming env has different value.

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18365:

Fix Version/s: (was: 1.11.1)
   1.11.2

> The same sql in a batch env and a streaming env has different value.
> 
>
> Key: FLINK-18365
> URL: https://issues.apache.org/jira/browse/FLINK-18365
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.11.0
>Reporter: xiaojin.wy
>Priority: Major
> Fix For: 1.11.2
>
>
> I use the sql-gateway to run this sql.
> *The input table is:*
>  CREATE TABLE `scott_dept` (
>   deptno INT,
>   dname VARCHAR,
>   loc VARCHAR
> ) WITH (
>   'format.field-delimiter'='|',
>   'connector.type'='filesystem',
>   'format.derive-schema'='true',
>   
> 'connector.path'='/defender_test_data/daily_regression_stream_blink_sql_1.10/test_scalar/sources/scott_dept.csv',
>   'format.type'='csv'
> )
> *The input data is:*
> 10|ACCOUNTING|NEW YORK
> 20|RESEARCH|DALLAS
> 30|SALES|CHICAGO
> 40|OPERATIONS|BOSTON
> *The sql is :*
> select deptno, (select count(*) from scott_emp where 1 = 0) as x from 
> scott_dept
> *The error:*
> In a batch environment, the result value is:10|0\n20|0\n30|0\n40|0
> In a streaming environment, the result value 
> is:10|None\n20|None\n30|None\n40|None



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16622) Type ARRAY of table field 'id' does not match with the physical type LEGACY('ARRAY', 'ANY<[Ljava.lang.String;, rO0ABXNyAD......>') of the 'id' field of the Table

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16622?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-16622:

Fix Version/s: (was: 1.11.1)
   1.11.2

> Type ARRAY of table field 'id' does not match with the physical type 
> LEGACY('ARRAY', 'ANY<[Ljava.lang.String;, rO0ABXNyAD..>') of the 'id' 
> field of the TableSource return type.
> 
>
> Key: FLINK-16622
> URL: https://issues.apache.org/jira/browse/FLINK-16622
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.10.0
>Reporter: gkgkgk
>Assignee: Jark Wu
>Priority: Major
> Fix For: 1.10.2, 1.11.2
>
>
> when i define a ARRAY field in table schema and define field in 
> 'format.json-schema' like this 
> 'format.json-schema' = '{
> "type": "object",
> "properties": {
>   "id": {
>   "type": "array",
>   "items":{"type":"string"}
>   }
>   }'
> there is a error flow this:
> Exception in thread "main" org.apache.flink.table.api.ValidationException: 
> Type ARRAY of table field 'id' does not match with the physical type 
> LEGACY('ARRAY', 'ANY<[Ljava.lang.String;, rO0ABXNyADdvc..>') of the 'id' 
> field of the TableSource return type.
> the flow is my ddl:
> --source
> CREATE TABLE dwd_user_log (
>   id ARRAY,
>   ctime TIMESTAMP(3),
>   pageId VARCHAR,
>   deviceId VARCHAR,
>   WATERMARK FOR ctime AS ctime - INTERVAL '10' SECOND
> ) WITH (
>   'connector.type' = 'kafka',
>   'connector.version' = 'universal',
>   'connector.topic' = 'dev_dwd_user_log_02',
>   'connector.startup-mode' = 'latest-offset',
>   'connector.properties.0.key' = 'zookeeper.connect',
>   'connector.properties.0.value' = 'node14.example.com:2181',
>   'connector.properties.1.key' = 'bootstrap.servers',
>   'connector.properties.1.value' = 'node14.example.com:9092',
>   'connector.properties.2.key' = 'group.id',
>   'connector.properties.2.value' = 'dev-group',
>   'update-mode' = 'append',
>   'format.type' = 'json',
>   -- 'format.derive-schema' = 'true'
>   'format.json-schema' = '{
> "type": "object",
> "properties": {
>   "id": {
>   "type": "array",
>   "items":{"type":"string"}
>   },
>   "ctime": {
>   "type": "string",
>   "format": "date-time"
>   },
>   "pageId": {
>   "type": "string"
>   },
>   "deviceId": {
>   "type": "string"
>   }
> }
>   }'
>   -- 'schema.1.rowtime.timestamps.type' = 'from-field',
>   -- 'schema.1.rowtime.timestamps.from' = 'ctime',
>   -- 'schema.1.rowtime.watermarks.type' = 'periodic-bounded',
>   -- 'schema.1.rowtime.watermarks.delay' = '1'
>   -- 'schema.1.from' = 'ctime'
> );  
> -- sink
> -- sink for pv
> CREATE TABLE dws_pv (
>   windowStart TIMESTAMP(3),
>   windowEnd TIMESTAMP(3),
>   pageId VARCHAR,
>   id ARRAY,
>   viewCount BIGINT
> ) WITH (
>   'connector.type' = 'kafka',
>   'connector.version' = 'universal',
>   'connector.topic' = 'dev_dws_pvuv_02',
>   'connector.startup-mode' = 'latest-offset',
>   'connector.properties.0.key' = 'zookeeper.connect',
>   'connector.properties.0.value' = 'node14.example.com:2181',
>   'connector.properties.1.key' = 'bootstrap.servers',
>   'connector.properties.1.value' = 'node14.example.com:9092',
>   'connector.properties.2.key' = 'group.id',
>   'connector.properties.2.value' = 'dev-group',
>   'update-mode' = 'append',
>   'format.type' = 'json',
>   'format.derive-schema' = 'true'
> );
> -- pv
> INSERT INTO dws_pv
> SELECT
>   TUMBLE_START(ctime, INTERVAL '20' SECOND)  AS windowStart,
>   TUMBLE_END(ctime, INTERVAL '20' SECOND)  AS windowEnd,
>   pageId,
>   id,
>   COUNT(deviceId) AS viewCount
> FROM dwd_user_log
> GROUP BY TUMBLE(ctime, INTERVAL '20' SECOND),pageId,id;



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17964) Hivemodule does not support map type

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17964?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-17964:

Fix Version/s: (was: 1.11.1)
   1.11.2

> Hivemodule does not support map type
> 
>
> Key: FLINK-17964
> URL: https://issues.apache.org/jira/browse/FLINK-17964
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.10.1
>Reporter: Shuai Xia
>Priority: Major
> Fix For: 1.12.0, 1.11.2
>
>
> {code:java}
> //代码占位符
> Exception in thread "main" scala.MatchError: MAP (of class 
> org.apache.flink.table.types.logical.LogicalTypeRoot)Exception in thread 
> "main" scala.MatchError: MAP (of class 
> org.apache.flink.table.types.logical.LogicalTypeRoot) at 
> org.apache.flink.table.planner.codegen.CodeGenUtils$.hashCodeForType(CodeGenUtils.scala:212)
>  at 
> org.apache.flink.table.planner.codegen.HashCodeGenerator$$anonfun$2.apply(HashCodeGenerator.scala:97)
>  at 
> org.apache.flink.table.planner.codegen.HashCodeGenerator$$anonfun$2.apply(HashCodeGenerator.scala:91)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
>  at 
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>  at scala.collection.mutable.WrappedArray.foreach(WrappedArray.scala:35) at 
> scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at 
> scala.collection.AbstractTraversable.map(Traversable.scala:104) at 
> org.apache.flink.table.planner.codegen.HashCodeGenerator$.generateCodeBody(HashCodeGenerator.scala:91)
>  at 
> org.apache.flink.table.planner.codegen.HashCodeGenerator$.generateRowHash(HashCodeGenerator.scala:61)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.batch.BatchExecExchange.translateToPlanInternal(BatchExecExchange.scala:182)
>  at 
> org.apache.flink.table.planner.plan.nodes.physical.batch.BatchExecExchange.translateToPlanInternal(BatchExecExchange.scala:52)
>  at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode$class.translateToPlan(ExecNode.scala:58)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18581) Cannot find GC cleaner with java version previous 8u202

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18581:

Fix Version/s: (was: 1.11.1)
   1.11.2

> Cannot find GC cleaner with java version previous 8u202
> ---
>
> Key: FLINK-18581
> URL: https://issues.apache.org/jira/browse/FLINK-18581
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.11.0
>Reporter: Xintong Song
>Priority: Critical
> Fix For: 1.12.0, 1.11.2
>
>
> {{JavaGcCleanerWrapper}} is looking for the package-private method 
> {{Reference.tryHandlePending}} using reflection. However, the method is first 
> introduced in the version 8u202. Therefore, if an older version JDK is used, 
> the method cannot be found and Flink will fail.
> See also this [ML 
> thread|http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Error-GC-Cleaner-Provider-Flink-1-11-0-td36565.html].



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18597) elasticsearch sql insert function cannot set documentid

2020-07-14 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157849#comment-17157849
 ] 

Jark Wu commented on FLINK-18597:
-

If you want to specify document-id, please refer to the docs: 
https://ci.apache.org/projects/flink/flink-docs-master/dev/table/connectors/elasticsearch.html#key-handling

 Lastly, I'd suggest to reach out to the mailing list first. I'd suggest 
creating an issue only when you have fully analyzed a problem.

> elasticsearch sql insert function cannot set documentid
> ---
>
> Key: FLINK-18597
> URL: https://issues.apache.org/jira/browse/FLINK-18597
> Project: Flink
>  Issue Type: Test
>Affects Versions: 1.11.0
>Reporter: xy
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18449) Make topic discovery and partition discovery configurable for FlinkKafkaConsumer in Table API

2020-07-14 Thread Shengkai Fang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157848#comment-17157848
 ] 

Shengkai Fang commented on FLINK-18449:
---

Hi, all. I find that FlinkKafkaProducer doesn't support partition discovery 
now. So it's may be more clear to rename 
{{'topic-partition-discovery.interval'}} to 
{{'scan.topic-partition-discovery.interval'}}. Considering Jark's proposal, I 
think the following options is better,
{code:java}
'topic' = 'topic'// for single topic
'topic' = 'topic-1, topic-2,..., topic-n',   // for list topic
'topic-pattern' = 'topic*'   // for use of regex,  "topic" 
and "topic-pattern" can't coexist
'scan.topic-partition-discovery.interval' = '5s'  // for both topic discovery
{code}

> Make topic discovery and partition discovery configurable for 
> FlinkKafkaConsumer in Table API
> -
>
> Key: FLINK-18449
> URL: https://issues.apache.org/jira/browse/FLINK-18449
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka, Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Shengkai Fang
>Assignee: Shengkai Fang
>Priority: Major
> Fix For: 1.12.0
>
>
> In streaming api, we can use regex to find topic and enable partiton 
> discovery by setting non-negative value for 
> `{{flink.partition-discovery.interval-millis}}`. However, it's not work in 
> table api. I think we can add options such as 'topic-regex' and 
> '{{partition-discovery.interval-millis}}' in WITH block for users.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12903: [FLINK-18492][python] Extract the implementation logic of Beam in Operations

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12903:
URL: https://github.com/apache/flink/pull/12903#issuecomment-658519879


   
   ## CI report:
   
   * aa4bb627890b6e46ef36f44528a24aa304bb7064 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4521)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18597) elasticsearch sql insert function cannot set documentid

2020-07-14 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18597?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu updated FLINK-18597:

Priority: Minor  (was: Critical)

> elasticsearch sql insert function cannot set documentid
> ---
>
> Key: FLINK-18597
> URL: https://issues.apache.org/jira/browse/FLINK-18597
> Project: Flink
>  Issue Type: Test
>Affects Versions: 1.11.0
>Reporter: xy
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] godfreyhe commented on a change in pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


godfreyhe commented on a change in pull request #12866:
URL: https://github.com/apache/flink/pull/12866#discussion_r454768224



##
File path: 
flink-table/flink-table-planner-blink/src/test/java/org/apache/flink/table/planner/factories/TestValuesTableFactory.java
##
@@ -540,40 +541,55 @@ private boolean 
isRetainedAfterApplyingFilterPredicates(Row row) {
private boolean binaryFilterApplies(CallExpression binExpr, Row 
row) {
List children = binExpr.getChildren();
Preconditions.checkArgument(children.size() == 2);
-   Comparable lhsValue = getValue(children.get(0), row);
-   Comparable rhsValue = getValue(children.get(1), row);
-   FunctionDefinition functionDefinition = 
binExpr.getFunctionDefinition();
+   Object lhsValue = getValue(children.get(0), row);
+   Object rhsValue = getValue(children.get(1), row);
+   // validate that literal is comparable
+   if (!isComparable(lhsValue, binExpr) || 
!isComparable(rhsValue, binExpr)) {

Review comment:
   how about check whether the type of a field is comparable in 
`shouldPushDownUnaryExpression` method ? then the logic of 
`binaryFilterApplies` need not to change





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18601) TaskManager shut down often in standalone cluster

2020-07-14 Thread Echo Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18601?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Echo Lee updated FLINK-18601:
-
Summary: TaskManager shut down often in standalone cluster  (was: 
Taskmanager shut down often in standalone cluster)

> TaskManager shut down often in standalone cluster
> -
>
> Key: FLINK-18601
> URL: https://issues.apache.org/jira/browse/FLINK-18601
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Echo Lee
>Priority: Major
>
> The situation we encountered is that the job is running on the standalone 
> cluster. Cancelling the job occasionally causes the taskmanager to shut down. 
> I am not sure whether this is a problem. and Some of the logs are as follows:
> {code:java}
> 2020-07-14 20:16:00.169 [Cancellation Watchdog for KeyedProcess -> (Filter -> 
> Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] ERROR 
> org.apache.flink.runtime.taskexecutor.TaskExecutor  - Task did not exit 
> gracefully within 180 + seconds.
> org.apache.flink.util.FlinkRuntimeException: Task did not exit gracefully 
> within 180 + seconds.
> at 
> org.apache.flink.runtime.taskmanager.Task$TaskCancelerWatchDog.run(Task.java:1572)
> at java.base/java.lang.Thread.run(Thread.java:834)
> 2020-07-14 20:16:00.169 [Cancellation Watchdog for KeyedProcess -> (Filter -> 
> Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] ERROR 
> org.apache.flink.runtime.taskexecutor.TaskManagerRunner  - Fatal error 
> occurred while executing the TaskManager. Shutting it down...
> org.apache.flink.util.FlinkRuntimeException: Task did not exit gracefully 
> within 180 + seconds.
> at 
> org.apache.flink.runtime.taskmanager.Task$TaskCancelerWatchDog.run(Task.java:1572)
> at java.base/java.lang.Thread.run(Thread.java:834)
> 2020-07-14 20:16:00.170 [flink-akka.actor.default-dispatcher-27] INFO  
> org.apache.flink.runtime.taskexecutor.TaskExecutor  - Stopping TaskExecutor 
> akka.tcp://flink@10.3.67.116:35867/user/rpc/taskmanager_0.
> 2020-07-14 20:16:00.170 [flink-akka.actor.default-dispatcher-27] INFO  
> org.apache.flink.runtime.taskexecutor.TaskExecutor  - Close ResourceManager 
> connection fb5cf1580e59ff78505facc7010b099c.
> 2020-07-14 20:16:00.195 [Canceler/Interrupts for KeyedProcess -> (Filter -> 
> Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] WARN  
> org.apache.flink.runtime.taskmanager.Task  - Task 'KeyedProcess -> (Filter -> 
> Map, Sink: FeatureSinkToES) (1/1)' did not react to cancelling signal for 30 
> seconds, but is stuck in method:
>  
> org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.snapshotState(ElasticsearchSinkBase.java:325)
> app//org.apache.flink.streaming.util.functions.StreamingFunctionUtils.trySnapshotFunctionState(StreamingFunctionUtils.java:120)
> app//org.apache.flink.streaming.util.functions.StreamingFunctionUtils.snapshotFunctionState(StreamingFunctionUtils.java:101)
> app//org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.snapshotState(AbstractUdfStreamOperator.java:90)
> app//org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.snapshotState(StreamOperatorStateHandler.java:186)
> app//org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.snapshotState(StreamOperatorStateHandler.java:156)
> app//org.apache.flink.streaming.api.operators.AbstractStreamOperator.snapshotState(AbstractStreamOperator.java:314)
> app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.checkpointStreamOperator(SubtaskCheckpointCoordinatorImpl.java:614)
> app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.buildOperatorSnapshotFutures(SubtaskCheckpointCoordinatorImpl.java:540)
> app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.takeSnapshotSync(SubtaskCheckpointCoordinatorImpl.java:507)
> app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.checkpointState(SubtaskCheckpointCoordinatorImpl.java:266)
> app//org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$performCheckpoint$5(StreamTask.java:892)
> app//org.apache.flink.streaming.runtime.tasks.StreamTask$$Lambda$3269/0x7f3f8a9ca560.run(Unknown
>  Source)
> app//org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.runThrowing(StreamTaskActionExecutor.java:47)
> app//org.apache.flink.streaming.runtime.tasks.StreamTask.performCheckpoint(StreamTask.java:882)
> app//org.apache.flink.streaming.runtime.tasks.StreamTask.triggerCheckpointOnBarrier(StreamTask.java:850)
> app//org.apache.flink.streaming.runtime.io.CheckpointBarrierHandler.notifyCheckpoint(CheckpointBarrierHandler.java:113)
> 

[jira] [Created] (FLINK-18601) Taskmanager shut down often in standalone cluster

2020-07-14 Thread Echo Lee (Jira)
Echo Lee created FLINK-18601:


 Summary: Taskmanager shut down often in standalone cluster
 Key: FLINK-18601
 URL: https://issues.apache.org/jira/browse/FLINK-18601
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Task
Affects Versions: 1.11.0, 1.10.0
Reporter: Echo Lee


The situation we encountered is that the job is running on the standalone 
cluster. Cancelling the job occasionally causes the taskmanager to shut down. I 
am not sure whether this is a problem. and Some of the logs are as follows:
{code:java}
2020-07-14 20:16:00.169 [Cancellation Watchdog for KeyedProcess -> (Filter -> 
Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] ERROR 
org.apache.flink.runtime.taskexecutor.TaskExecutor  - Task did not exit 
gracefully within 180 + seconds.
org.apache.flink.util.FlinkRuntimeException: Task did not exit gracefully 
within 180 + seconds.
at 
org.apache.flink.runtime.taskmanager.Task$TaskCancelerWatchDog.run(Task.java:1572)
at java.base/java.lang.Thread.run(Thread.java:834)
2020-07-14 20:16:00.169 [Cancellation Watchdog for KeyedProcess -> (Filter -> 
Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] ERROR 
org.apache.flink.runtime.taskexecutor.TaskManagerRunner  - Fatal error occurred 
while executing the TaskManager. Shutting it down...
org.apache.flink.util.FlinkRuntimeException: Task did not exit gracefully 
within 180 + seconds.
at 
org.apache.flink.runtime.taskmanager.Task$TaskCancelerWatchDog.run(Task.java:1572)
at java.base/java.lang.Thread.run(Thread.java:834)
2020-07-14 20:16:00.170 [flink-akka.actor.default-dispatcher-27] INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor  - Stopping TaskExecutor 
akka.tcp://flink@10.3.67.116:35867/user/rpc/taskmanager_0.
2020-07-14 20:16:00.170 [flink-akka.actor.default-dispatcher-27] INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor  - Close ResourceManager 
connection fb5cf1580e59ff78505facc7010b099c.
2020-07-14 20:16:00.195 [Canceler/Interrupts for KeyedProcess -> (Filter -> 
Map, Sink: FeatureSinkToES) (1/1) (8c10a884bf480714460860f9d0e6131e).] WARN  
org.apache.flink.runtime.taskmanager.Task  - Task 'KeyedProcess -> (Filter -> 
Map, Sink: FeatureSinkToES) (1/1)' did not react to cancelling signal for 30 
seconds, but is stuck in method:
 
org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.snapshotState(ElasticsearchSinkBase.java:325)
app//org.apache.flink.streaming.util.functions.StreamingFunctionUtils.trySnapshotFunctionState(StreamingFunctionUtils.java:120)
app//org.apache.flink.streaming.util.functions.StreamingFunctionUtils.snapshotFunctionState(StreamingFunctionUtils.java:101)
app//org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.snapshotState(AbstractUdfStreamOperator.java:90)
app//org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.snapshotState(StreamOperatorStateHandler.java:186)
app//org.apache.flink.streaming.api.operators.StreamOperatorStateHandler.snapshotState(StreamOperatorStateHandler.java:156)
app//org.apache.flink.streaming.api.operators.AbstractStreamOperator.snapshotState(AbstractStreamOperator.java:314)
app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.checkpointStreamOperator(SubtaskCheckpointCoordinatorImpl.java:614)
app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.buildOperatorSnapshotFutures(SubtaskCheckpointCoordinatorImpl.java:540)
app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.takeSnapshotSync(SubtaskCheckpointCoordinatorImpl.java:507)
app//org.apache.flink.streaming.runtime.tasks.SubtaskCheckpointCoordinatorImpl.checkpointState(SubtaskCheckpointCoordinatorImpl.java:266)
app//org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$performCheckpoint$5(StreamTask.java:892)
app//org.apache.flink.streaming.runtime.tasks.StreamTask$$Lambda$3269/0x7f3f8a9ca560.run(Unknown
 Source)
app//org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.runThrowing(StreamTaskActionExecutor.java:47)
app//org.apache.flink.streaming.runtime.tasks.StreamTask.performCheckpoint(StreamTask.java:882)
app//org.apache.flink.streaming.runtime.tasks.StreamTask.triggerCheckpointOnBarrier(StreamTask.java:850)
app//org.apache.flink.streaming.runtime.io.CheckpointBarrierHandler.notifyCheckpoint(CheckpointBarrierHandler.java:113)
app//org.apache.flink.streaming.runtime.io.CheckpointBarrierAligner.processBarrier(CheckpointBarrierAligner.java:137)
app//org.apache.flink.streaming.runtime.io.CheckpointedInputGate.pollNext(CheckpointedInputGate.java:93)
app//org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:158)
app//org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:67)

[jira] [Commented] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157844#comment-17157844
 ] 

Dian Fu commented on FLINK-18600:
-

Not sure the reason why we choose oracle JDK and if we could switch to Open JDK 
for the tests?

> Kerberized YARN per-job on Docker test failed to download JDK 8u251
> ---
>
> Key: FLINK-18600
> URL: https://issues.apache.org/jira/browse/FLINK-18600
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dian Fu
>Priority: Blocker
> Fix For: 1.12.0, 1.11.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> + mkdir -p /usr/java/default
> + curl -Ls 
> https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
>  -H Cookie: oraclelicense=accept-securebackup-cookie
> + tar --strip-components=1 -xz -C /usr/java/default/
> gzip: stdin: not in gzip format
> tar: Child returned status 1
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18600:

Priority: Blocker  (was: Critical)

> Kerberized YARN per-job on Docker test failed to download JDK 8u251
> ---
>
> Key: FLINK-18600
> URL: https://issues.apache.org/jira/browse/FLINK-18600
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dian Fu
>Priority: Blocker
> Fix For: 1.12.0, 1.11.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> + mkdir -p /usr/java/default
> + curl -Ls 
> https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
>  -H Cookie: oraclelicense=accept-securebackup-cookie
> + tar --strip-components=1 -xz -C /usr/java/default/
> gzip: stdin: not in gzip format
> tar: Child returned status 1
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157843#comment-17157843
 ] 

Dian Fu commented on FLINK-18600:
-

I have checked that I can download the version *8u251* through the oracle 
website after accepting the license. However, it's not available if download 
through CLI using the same command. I'm wondering if it's because the cookie 
has changed or some other reason?

> Kerberized YARN per-job on Docker test failed to download JDK 8u251
> ---
>
> Key: FLINK-18600
> URL: https://issues.apache.org/jira/browse/FLINK-18600
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dian Fu
>Priority: Critical
> Fix For: 1.12.0, 1.11.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> + mkdir -p /usr/java/default
> + curl -Ls 
> https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
>  -H Cookie: oraclelicense=accept-securebackup-cookie
> + tar --strip-components=1 -xz -C /usr/java/default/
> gzip: stdin: not in gzip format
> tar: Child returned status 1
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18540) Unnecessary retract messages when setIdleStateRetentionTime before converting dynamic table to retract stream

2020-07-14 Thread WeiqiangSun (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157838#comment-17157838
 ] 

WeiqiangSun commented on FLINK-18540:
-

[~libenchao] Mini batch may not work well with retract messages for HBase 
connector. Unnecessary retract messages can cause  puts and deletes with same 
timestamp. As for HBase, deletes can mask puts with the same timestamp.

 https://issues.apache.org/jira/browse/HBASE-2256 
https://issues.apache.org/jira/browse/HBASE-8721

> Unnecessary retract messages when setIdleStateRetentionTime before converting 
> dynamic table to retract stream
> -
>
> Key: FLINK-18540
> URL: https://issues.apache.org/jira/browse/FLINK-18540
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner, Table SQL / Runtime
>Affects Versions: 1.10.0
>Reporter: WeiqiangSun
>Priority: Critical
>
> Unnecessary retract messages when setIdleStateRetentionTime before converting 
> dynamic table to retract stream。  
>  
> {code:java}
> final StreamExecutionEnvironment env = 
> StreamExecutionEnvironment.getExecutionEnvironment();
> env.setParallelism(1);
> EnvironmentSettings es = 
> EnvironmentSettings.newInstance().inStreamingMode().build();
> StreamTableEnvironment fsTableEnv = StreamTableEnvironment.create(env,es);
> DataStream ds = env.fromElements(5,4,3,2,1);
> // if do not setIdleStateRetentionTime, it will output only one message as 
> expected like '(true,5)' 
> // if do setIdleStateRetentionTime, it will output unnecessary retract 
> messages like '(false,5)/(true,5)'
> fsTableEnv.getConfig().setIdleStateRetentionTime(Time.hours(100),Time.hours(1000));
> fsTableEnv.createTemporaryView("test",ds,"id");
> Table test = fsTableEnv.sqlQuery("select max(id) from test");
> fsTableEnv.toRetractStream(test,Row.class).printToErr();
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12903: [FLINK-18492][python] Extract the implementation logic of Beam in Operations

2020-07-14 Thread GitBox


flinkbot commented on pull request #12903:
URL: https://github.com/apache/flink/pull/12903#issuecomment-658519879


   
   ## CI report:
   
   * aa4bb627890b6e46ef36f44528a24aa304bb7064 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18600?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157832#comment-17157832
 ] 

Jark Wu commented on FLINK-18600:
-

Looks like we started to use this JDK link since FLINK-18485. Maybe this link 
is not stable. 

cc [~dwysakowicz]

> Kerberized YARN per-job on Docker test failed to download JDK 8u251
> ---
>
> Key: FLINK-18600
> URL: https://issues.apache.org/jira/browse/FLINK-18600
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dian Fu
>Priority: Critical
> Fix For: 1.12.0, 1.11.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> + mkdir -p /usr/java/default
> + curl -Ls 
> https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
>  -H Cookie: oraclelicense=accept-securebackup-cookie
> + tar --strip-components=1 -xz -C /usr/java/default/
> gzip: stdin: not in gzip format
> tar: Child returned status 1
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12902: [FLINK-18415][python] Support TableResult#collect in the Python Table…

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12902:
URL: https://github.com/apache/flink/pull/12902#issuecomment-658509076


   
   ## CI report:
   
   * 802dfd4a0a2fb3e1b6e9a31fbc850c18f00c5c2b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4518)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12866:
URL: https://github.com/apache/flink/pull/12866#issuecomment-656576219


   
   ## CI report:
   
   * 02c81cff07ef76dc20a7c618112cc7a172856720 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4497)
 
   * 5eb302884e390b6099859abcd7955e7f1c4e034e Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4517)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] lirui-apache commented on pull request #12888: [FLINK-18588] hive ddl create table support 'if not exists'

2020-07-14 Thread GitBox


lirui-apache commented on pull request #12888:
URL: https://github.com/apache/flink/pull/12888#issuecomment-658513513


   Thanks @wtog for the contribution. Could we also have some test for this in 
`HiveDialectITCase`?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12903: [FLINK-18492][python] Extract the implementation logic of Beam in Operations

2020-07-14 Thread GitBox


flinkbot commented on pull request #12903:
URL: https://github.com/apache/flink/pull/12903#issuecomment-658512879


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit aa4bb627890b6e46ef36f44528a24aa304bb7064 (Wed Jul 15 
02:45:44 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] HuangXingBo opened a new pull request #12903: [FLINK-18492][python] Extract the implementation logic of Beam in Operations

2020-07-14 Thread GitBox


HuangXingBo opened a new pull request #12903:
URL: https://github.com/apache/flink/pull/12903


   ## What is the purpose of the change
   
   *This pull request extract the implementation logic of Beam in Operations*
   
   
   ## Brief change log
   
 - *Extract the beam related logic of fast_operations to beam_operations*
 - *rename operations.py to beam_slow_operations.py*
   
   ## Verifying this change
*current tests are enough*
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18492) Extract the implementation logic of Beam in Operations

2020-07-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18492?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18492:
---
Labels: pull-request-available  (was: )

> Extract the implementation logic of Beam in Operations
> --
>
> Key: FLINK-18492
> URL: https://issues.apache.org/jira/browse/FLINK-18492
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> Extract the implementation logic of Beam in Operations, so that the 
> implementation of general operations and Beam operations can be decoupled



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17636) SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read from released RecoveredInputChannel

2020-07-14 Thread Zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijiang updated FLINK-17636:
-
Affects Version/s: 1.11.0

> SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read 
> from released RecoveredInputChannel
> -
>
> Key: FLINK-17636
> URL: https://issues.apache.org/jira/browse/FLINK-17636
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Assignee: Zhijiang
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1080=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=4ed44b66-cdd6-5dcf-5f6a-88b07dda665d
> {code}
> 2020-05-12T11:39:28.7058732Z [ERROR] Tests run: 22, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 2.643 s <<< FAILURE! - in 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest
> 2020-05-12T11:39:28.7066377Z [ERROR] 
> testConcurrentReadStateAndProcessAndClose(org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest)
>   Time elapsed: 0.032 s  <<< ERROR!
> 2020-05-12T11:39:28.7067491Z java.util.concurrent.ExecutionException: 
> java.lang.IllegalStateException: Trying to read from released 
> RecoveredInputChannel
> 2020-05-12T11:39:28.7068238Z  at 
> java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 2020-05-12T11:39:28.7068795Z  at 
> java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 2020-05-12T11:39:28.7069538Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.RemoteInputChannelTest.submitTasksAndWaitForResults(RemoteInputChannelTest.java:1228)
> 2020-05-12T11:39:28.7070595Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest.testConcurrentReadStateAndProcessAndClose(SingleInputGateTest.java:235)
> 2020-05-12T11:39:28.7075974Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-12T11:39:28.7076784Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-12T11:39:28.7077522Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-12T11:39:28.7078212Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-12T11:39:28.7078846Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-12T11:39:28.7079607Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-12T11:39:28.7080383Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-12T11:39:28.7081173Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-12T11:39:28.7081937Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-12T11:39:28.7082708Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-12T11:39:28.7083422Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-12T11:39:28.7084148Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-12T11:39:28.7084933Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-12T11:39:28.7085562Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-12T11:39:28.7086162Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-12T11:39:28.7086806Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-12T11:39:28.7087434Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-12T11:39:28.7088036Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-12T11:39:28.7088647Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> 2020-05-12T11:39:28.7089328Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> 2020-05-12T11:39:28.7090106Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> 2020-05-12T11:39:28.7090811Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> 2020-05-12T11:39:28.7091674Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> 2020-05-12T11:39:28.7102178Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
> 2020-05-12T11:39:28.7103048Z  at 
> 

[jira] [Updated] (FLINK-17636) SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read from released RecoveredInputChannel

2020-07-14 Thread Zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijiang updated FLINK-17636:
-
Fix Version/s: 1.11.0

> SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read 
> from released RecoveredInputChannel
> -
>
> Key: FLINK-17636
> URL: https://issues.apache.org/jira/browse/FLINK-17636
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Assignee: Zhijiang
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0, 1.12.0
>
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1080=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=4ed44b66-cdd6-5dcf-5f6a-88b07dda665d
> {code}
> 2020-05-12T11:39:28.7058732Z [ERROR] Tests run: 22, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 2.643 s <<< FAILURE! - in 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest
> 2020-05-12T11:39:28.7066377Z [ERROR] 
> testConcurrentReadStateAndProcessAndClose(org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest)
>   Time elapsed: 0.032 s  <<< ERROR!
> 2020-05-12T11:39:28.7067491Z java.util.concurrent.ExecutionException: 
> java.lang.IllegalStateException: Trying to read from released 
> RecoveredInputChannel
> 2020-05-12T11:39:28.7068238Z  at 
> java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 2020-05-12T11:39:28.7068795Z  at 
> java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 2020-05-12T11:39:28.7069538Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.RemoteInputChannelTest.submitTasksAndWaitForResults(RemoteInputChannelTest.java:1228)
> 2020-05-12T11:39:28.7070595Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest.testConcurrentReadStateAndProcessAndClose(SingleInputGateTest.java:235)
> 2020-05-12T11:39:28.7075974Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-12T11:39:28.7076784Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-12T11:39:28.7077522Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-12T11:39:28.7078212Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-12T11:39:28.7078846Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-12T11:39:28.7079607Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-12T11:39:28.7080383Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-12T11:39:28.7081173Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-12T11:39:28.7081937Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-12T11:39:28.7082708Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-12T11:39:28.7083422Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-12T11:39:28.7084148Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-12T11:39:28.7084933Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-12T11:39:28.7085562Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-12T11:39:28.7086162Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-12T11:39:28.7086806Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-12T11:39:28.7087434Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-12T11:39:28.7088036Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-12T11:39:28.7088647Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> 2020-05-12T11:39:28.7089328Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> 2020-05-12T11:39:28.7090106Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> 2020-05-12T11:39:28.7090811Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> 2020-05-12T11:39:28.7091674Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> 2020-05-12T11:39:28.7102178Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
> 2020-05-12T11:39:28.7103048Z  at 
> 

[jira] [Closed] (FLINK-17636) SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read from released RecoveredInputChannel

2020-07-14 Thread Zhijiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17636?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhijiang closed FLINK-17636.

Resolution: Fixed

Merged in release-1.11: 57e57624deded5594749cbccf17a5878c6dc81b8
Merged in master: a9cce5467b3f0628e665cd8c7dd9b2e5958f0fd7

> SingleInputGateTest.testConcurrentReadStateAndProcessAndClose: Trying to read 
> from released RecoveredInputChannel
> -
>
> Key: FLINK-17636
> URL: https://issues.apache.org/jira/browse/FLINK-17636
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Reporter: Robert Metzger
>Assignee: Zhijiang
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> CI: 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1080=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=4ed44b66-cdd6-5dcf-5f6a-88b07dda665d
> {code}
> 2020-05-12T11:39:28.7058732Z [ERROR] Tests run: 22, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 2.643 s <<< FAILURE! - in 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest
> 2020-05-12T11:39:28.7066377Z [ERROR] 
> testConcurrentReadStateAndProcessAndClose(org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest)
>   Time elapsed: 0.032 s  <<< ERROR!
> 2020-05-12T11:39:28.7067491Z java.util.concurrent.ExecutionException: 
> java.lang.IllegalStateException: Trying to read from released 
> RecoveredInputChannel
> 2020-05-12T11:39:28.7068238Z  at 
> java.util.concurrent.FutureTask.report(FutureTask.java:122)
> 2020-05-12T11:39:28.7068795Z  at 
> java.util.concurrent.FutureTask.get(FutureTask.java:192)
> 2020-05-12T11:39:28.7069538Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.RemoteInputChannelTest.submitTasksAndWaitForResults(RemoteInputChannelTest.java:1228)
> 2020-05-12T11:39:28.7070595Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGateTest.testConcurrentReadStateAndProcessAndClose(SingleInputGateTest.java:235)
> 2020-05-12T11:39:28.7075974Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-12T11:39:28.7076784Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-12T11:39:28.7077522Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-12T11:39:28.7078212Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-12T11:39:28.7078846Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-12T11:39:28.7079607Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-12T11:39:28.7080383Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-12T11:39:28.7081173Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-12T11:39:28.7081937Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
> 2020-05-12T11:39:28.7082708Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-05-12T11:39:28.7083422Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-05-12T11:39:28.7084148Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-05-12T11:39:28.7084933Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-05-12T11:39:28.7085562Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-05-12T11:39:28.7086162Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-05-12T11:39:28.7086806Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-05-12T11:39:28.7087434Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-05-12T11:39:28.7088036Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-05-12T11:39:28.7088647Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> 2020-05-12T11:39:28.7089328Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> 2020-05-12T11:39:28.7090106Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> 2020-05-12T11:39:28.7090811Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> 2020-05-12T11:39:28.7091674Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> 2020-05-12T11:39:28.7102178Z  at 
> 

[GitHub] [flink] zhijiangW merged pull request #12879: [FLINK-17636][tests] Fix unstable test SingleInputGateTest#testConcur…

2020-07-14 Thread GitBox


zhijiangW merged pull request #12879:
URL: https://github.com/apache/flink/pull/12879


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhijiangW merged pull request #12898: [FLINK-17636][tests] Fix unstable test SingleInputGateTest#testConcurrentReadStateAndProcessAndClose

2020-07-14 Thread GitBox


zhijiangW merged pull request #12898:
URL: https://github.com/apache/flink/pull/12898


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18599) Compile error when use windowAll and TumblingProcessingTimeWindows

2020-07-14 Thread henvealf (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

henvealf updated FLINK-18599:
-
Description: 
Code:
{code:java}
import org.apache.commons.lang3.StringUtils

import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment
import 
org.apache.flink.streaming.api.windowing.assigners.{TumblingProcessingTimeWindows}
import org.apache.flink.streaming.api.windowing.time.Time

import org.apache.flink.streaming.api.scala._
   
val env = StreamExecutionEnvironment.getExecutionEnvironment
val stream = env.fromElements("a", "b", "c")

stream
  .filter((str: String) => StringUtils.isNotEmpty(str))
  .map( _ => 1)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
  .reduce((a1, a2) => a1 + a2)
  .print()

{code}
Compile failed:
{code:java}
error: type mismatch;
 found   : 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 required: org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[_ 
>: Int, ?]
Note: Object <: Any (and 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 <: 
org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[Object,org.apache.flink.streaming.api.windowing.windows.TimeWindow]),
 but Java-defined class WindowAssigner is invariant in type T.
You may wish to investigate a wildcard type such as `_ <: Any`. (SLS 3.2.10)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
 ^
one error found

{code}
 What went wrong?

 Scala version: 2.11

 Flink version: 1.11

 Thanks.

  was:
Code:
{code:java}
import org.apache.commons.lang3.StringUtils

import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment
import 
org.apache.flink.streaming.api.windowing.assigners.{TumblingProcessingTimeWindows}
import org.apache.flink.streaming.api.windowing.time.Time

import org.apache.flink.streaming.api.scala._
   
val env = StreamExecutionEnvironment.getExecutionEnvironment
val stream = env.fromElements("a", "b", "c")

stream
  .filter((str: String) => StringUtils.isNotEmpty(str))
  .map( _ => 1)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
  .reduce((a1, a2) => a1 + a2)
  .print()

{code}
Compile failed:
{code:java}
error: type mismatch;
 found   : 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 required: org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[_ 
>: Int, ?]
Note: Object <: Any (and 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 <: 
org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[Object,org.apache.flink.streaming.api.windowing.windows.TimeWindow]),
 but Java-defined class WindowAssigner is invariant in type T.
You may wish to investigate a wildcard type such as `_ <: Any`. (SLS 3.2.10)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
 ^
one error found

{code}
 What went wrong?

 Thanks.


> Compile error when use windowAll and TumblingProcessingTimeWindows
> --
>
> Key: FLINK-18599
> URL: https://issues.apache.org/jira/browse/FLINK-18599
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.11.0
>Reporter: henvealf
>Priority: Major
>
> Code:
> {code:java}
> import org.apache.commons.lang3.StringUtils
> import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment
> import 
> org.apache.flink.streaming.api.windowing.assigners.{TumblingProcessingTimeWindows}
> import org.apache.flink.streaming.api.windowing.time.Time
> import org.apache.flink.streaming.api.scala._
>
> val env = StreamExecutionEnvironment.getExecutionEnvironment
> val stream = env.fromElements("a", "b", "c")
> stream
>   .filter((str: String) => StringUtils.isNotEmpty(str))
>   .map( _ => 1)
>   .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
>   .reduce((a1, a2) => a1 + a2)
>   .print()
> {code}
> Compile failed:
> {code:java}
> error: type mismatch;
>  found   : 
> org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
>  required: 
> org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[_ >: Int, ?]
> Note: Object <: Any (and 
> org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
>  <: 
> org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[Object,org.apache.flink.streaming.api.windowing.windows.TimeWindow]),
>  but Java-defined class WindowAssigner is invariant in type T.
> You may wish to investigate a wildcard type such as `_ <: Any`. (SLS 3.2.10)
>   

[jira] [Updated] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-18600:

Description: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
{code}
+ mkdir -p /usr/java/default
+ curl -Ls 
https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
 -H Cookie: oraclelicense=accept-securebackup-cookie
+ tar --strip-components=1 -xz -C /usr/java/default/

gzip: stdin: not in gzip format
tar: Child returned status 1
{code}

> Kerberized YARN per-job on Docker test failed to download JDK 8u251
> ---
>
> Key: FLINK-18600
> URL: https://issues.apache.org/jira/browse/FLINK-18600
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dian Fu
>Priority: Critical
> Fix For: 1.12.0, 1.11.2
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=4514=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> + mkdir -p /usr/java/default
> + curl -Ls 
> https://download.oracle.com/otn-pub/java/jdk/8u251-b08/3d5a2bb8f8d4428bbe94aed7ec7ae784/jdk-8u251-linux-x64.tar.gz
>  -H Cookie: oraclelicense=accept-securebackup-cookie
> + tar --strip-components=1 -xz -C /usr/java/default/
> gzip: stdin: not in gzip format
> tar: Child returned status 1
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18600) Kerberized YARN per-job on Docker test failed to download JDK 8u251

2020-07-14 Thread Dian Fu (Jira)
Dian Fu created FLINK-18600:
---

 Summary: Kerberized YARN per-job on Docker test failed to download 
JDK 8u251
 Key: FLINK-18600
 URL: https://issues.apache.org/jira/browse/FLINK-18600
 Project: Flink
  Issue Type: Bug
  Components: Deployment / YARN, Tests
Affects Versions: 1.11.0, 1.12.0
Reporter: Dian Fu
 Fix For: 1.12.0, 1.11.2






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12902: [FLINK-18415][python] Support TableResult#collect in the Python Table…

2020-07-14 Thread GitBox


flinkbot commented on pull request #12902:
URL: https://github.com/apache/flink/pull/12902#issuecomment-658509076


   
   ## CI report:
   
   * 802dfd4a0a2fb3e1b6e9a31fbc850c18f00c5c2b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12866:
URL: https://github.com/apache/flink/pull/12866#issuecomment-656576219


   
   ## CI report:
   
   * 02c81cff07ef76dc20a7c618112cc7a172856720 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4497)
 
   * 5eb302884e390b6099859abcd7955e7f1c4e034e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18599) Compile error when use windowAll and TumblingProcessingTimeWindows

2020-07-14 Thread henvealf (Jira)
henvealf created FLINK-18599:


 Summary: Compile error when use windowAll and 
TumblingProcessingTimeWindows
 Key: FLINK-18599
 URL: https://issues.apache.org/jira/browse/FLINK-18599
 Project: Flink
  Issue Type: Bug
  Components: API / DataStream
Affects Versions: 1.11.0
Reporter: henvealf


Code:
{code:java}
import org.apache.commons.lang3.StringUtils

import org.apache.flink.streaming.api.scala.StreamExecutionEnvironment
import 
org.apache.flink.streaming.api.windowing.assigners.{TumblingProcessingTimeWindows}
import org.apache.flink.streaming.api.windowing.time.Time

import org.apache.flink.streaming.api.scala._
   
val env = StreamExecutionEnvironment.getExecutionEnvironment
val stream = env.fromElements("a", "b", "c")

stream
  .filter((str: String) => StringUtils.isNotEmpty(str))
  .map( _ => 1)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
  .reduce((a1, a2) => a1 + a2)
  .print()

{code}
Compile failed:
{code:java}
error: type mismatch;
 found   : 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 required: org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[_ 
>: Int, ?]
Note: Object <: Any (and 
org.apache.flink.streaming.api.windowing.assigners.TumblingProcessingTimeWindows
 <: 
org.apache.flink.streaming.api.windowing.assigners.WindowAssigner[Object,org.apache.flink.streaming.api.windowing.windows.TimeWindow]),
 but Java-defined class WindowAssigner is invariant in type T.
You may wish to investigate a wildcard type such as `_ <: Any`. (SLS 3.2.10)
  .windowAll(TumblingProcessingTimeWindows.of(Time.seconds(5)))
 ^
one error found

{code}
 What went wrong?

 Thanks.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] liuyongvs commented on pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


liuyongvs commented on pull request #12866:
URL: https://github.com/apache/flink/pull/12866#issuecomment-658506250


   Hi @godfreyhe . Yes, i find some type is not Comparable, such as Period int 
getValueAs(Class clazz).
   Thanks for you review in great details. And I learned a lot



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18598) Add instructions for asynchronous execute in PyFlink doc

2020-07-14 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-18598:


 Summary: Add instructions for asynchronous execute in PyFlink doc
 Key: FLINK-18598
 URL: https://issues.apache.org/jira/browse/FLINK-18598
 Project: Flink
  Issue Type: Improvement
  Components: API / Python, Documentation
Affects Versions: 1.11.0
Reporter: Huang Xingbo


Add instructions for asynchronous execute in PyFlink doc



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18597) elasticsearch sql insert function cannot set documentid

2020-07-14 Thread Yangze Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18597?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157811#comment-17157811
 ] 

Yangze Guo commented on FLINK-18597:


I think it is fixed by FLINK-18583. Could you take a look? If that is the case, 
please close this ticket.

> elasticsearch sql insert function cannot set documentid
> ---
>
> Key: FLINK-18597
> URL: https://issues.apache.org/jira/browse/FLINK-18597
> Project: Flink
>  Issue Type: Test
>Affects Versions: 1.11.0
>Reporter: xy
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18597) elasticsearch sql insert function cannot set documentid

2020-07-14 Thread xy (Jira)
xy created FLINK-18597:
--

 Summary: elasticsearch sql insert function cannot set documentid
 Key: FLINK-18597
 URL: https://issues.apache.org/jira/browse/FLINK-18597
 Project: Flink
  Issue Type: Test
Affects Versions: 1.11.0
Reporter: xy






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12902: [FLINK-18415][python] Support TableResult#collect in the Python Table…

2020-07-14 Thread GitBox


flinkbot commented on pull request #12902:
URL: https://github.com/apache/flink/pull/12902#issuecomment-658497220


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 802dfd4a0a2fb3e1b6e9a31fbc850c18f00c5c2b (Wed Jul 15 
01:44:55 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18415) Support TableResult#collect in the Python Table API to align with the Java Table API

2020-07-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18415?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18415:
---
Labels: pull-request-available  (was: )

> Support TableResult#collect in the Python Table API to align with the Java 
> Table API
> 
>
> Key: FLINK-18415
> URL: https://issues.apache.org/jira/browse/FLINK-18415
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python
>Reporter: Dian Fu
>Assignee: Shuiqiang Chen
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] shuiqiangchen opened a new pull request #12902: [FLINK-18415][python] Support TableResult#collect in the Python Table…

2020-07-14 Thread GitBox


shuiqiangchen opened a new pull request #12902:
URL: https://github.com/apache/flink/pull/12902


   
   
   ## What is the purpose of the change
   
   *Support TableResult#collect in the Python Table API to align with the Java 
Table API*
   
   
   ## Brief change log
   
 - * Add collect() method for TableResult*
 - *add a new class named ClosableIterator to iterate the result of 
TableResult*
 - *add a type util to get/transform data between java row object and 
python object*
   
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   
 - *Add an integration test to test the collect() in test_table_api.py*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): ( no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: ( no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (no)
 - The S3 file system connector: (no)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? ( no)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12901: [FLINK-17285] [API / Python, chinese-translation] Translate "Python Table API" page into Chinese (part 1)

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12901:
URL: https://github.com/apache/flink/pull/12901#issuecomment-658445990


   
   ## CI report:
   
   * 221cb71c264d05e43eb77b5e11d9fb4c8834e7da Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4516)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12901: [FLINK-17285] [API / Python, chinese-translation] Translate "Python Table API" page into Chinese (part 1)

2020-07-14 Thread GitBox


flinkbot commented on pull request #12901:
URL: https://github.com/apache/flink/pull/12901#issuecomment-658445990


   
   ## CI report:
   
   * 221cb71c264d05e43eb77b5e11d9fb4c8834e7da UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12901: [FLINK-17285] [API / Python, chinese-translation] Translate "Python Table API" page into Chinese (part 1)

2020-07-14 Thread GitBox


flinkbot commented on pull request #12901:
URL: https://github.com/apache/flink/pull/12901#issuecomment-658438478


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 221cb71c264d05e43eb77b5e11d9fb4c8834e7da (Tue Jul 14 
22:09:45 UTC 2020)
   
✅no warnings
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17285) Translate "Python Table API" page into Chinese

2020-07-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-17285:
---
Labels: pull-request-available  (was: )

> Translate "Python Table API" page into Chinese 
> ---
>
> Key: FLINK-17285
> URL: https://issues.apache.org/jira/browse/FLINK-17285
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Python, chinese-translation
>Reporter: Zixuan Rao
>Assignee: Zixuan Rao
>Priority: Major
>  Labels: pull-request-available
>   Original Estimate: 168h
>  Remaining Estimate: 168h
>
> This issue is to add Chinese translation to Python Table API. Untranslated 
> paragraphs currently in ```*-zh.md``` will be replaced with Chinese 
> translations. Translation will be done with ```Google Translate``` and manual 
> corrections. I plan to submit a pull request this week. Thanks. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] billyrrr opened a new pull request #12901: [FLINK-17285] [API / Python, chinese-translation] Translate "Python Table API" page into Chinese (part 1)

2020-07-14 Thread GitBox


billyrrr opened a new pull request #12901:
URL: https://github.com/apache/flink/pull/12901


   
   
   ## What is the purpose of the change
   
   This change is part of many that will translate `*.zh.md` docs for Python 
Table API to Chinese Simplified. Untranslated paragraphs currently in 
```*-zh.md``` will be replaced with Chinese translations. Translation will be 
done with ```Google Translate``` and manual corrections. NOTE: this change only 
changes part of the files, and intends to receive feedbacks. 
   
   
   ## Brief change log
   
   Translate files:
- Improve docs/dev/table/python/common_questions.zh.md
- Translate docs/dev/table/python/index.zh.md
- Translate docs/dev/table/python/python_udfs.zh.md 
   
   ## Verifying this change
   
   This change is already covered by existing tests, such as document build 
test (if exists). 
   
   NOTE: some files are translated from `*.md` and others from `*.zh.md`. 
Please verify that the translated text contains the latest changes, even if 
this commit is based on updated `*.md` or `*.zh.md` files. 
   
   NOTE: custom headers should be checked for compatibility and are not applied 
every where. 
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no 
 - If yes, how is the feature documented? N/A 
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12900:
URL: https://github.com/apache/flink/pull/12900#issuecomment-658213136


   
   ## CI report:
   
   * ceb895be265bd445b9d36bfc52cf8018a508c333 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4508)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-654978676


   
   ## CI report:
   
   * 8c8f594927a2b72c23f8619c1932219585755054 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4509)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18573) InfluxDB reporter cannot be loaded as plugin

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-18573:
-
Description: 
The service entry is in the wrong directory (service vs service*s* )

It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.

  was:
The service entry is in the wrong directory (service vs service*s*)

It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.


> InfluxDB reporter cannot be loaded as plugin
> 
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> The service entry is in the wrong directory (service vs service*s* )
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18573) InfluxDB reporter cannot be loaded as plugin

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-18573:
-
Description: 
The service entry is in the wrong directory (service vs service*s*)

It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.

  was:
It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.


> InfluxDB reporter cannot be loaded as plugin
> 
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> The service entry is in the wrong directory (service vs service*s*)
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18573) InfluxDB reporter cannot be loaded as plugin

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-18573:
-
Description: 
The service entry is in the wrong directory (service vs services)

It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.

  was:
The service entry is in the wrong directory (service vs service*s* )

It causes the error:

 
2020-07-13 09:08:46.146 [main] WARN 
org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
(org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be found 
for reporter influxdb. Available factories: 
[org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
org.apache.flink.metrics.statsd.StatsDReporterFactory, 
org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
org.apache.flink.metrics.jmx.JMXReporterFactory].
2020-07-13 09:08:46.149 [main] INFO 
org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
metrics in intervals of 60 SECONDS for reporter slf4j of type 
org.apache.flink.metrics.slf4j.Slf4jReporter.


> InfluxDB reporter cannot be loaded as plugin
> 
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> The service entry is in the wrong directory (service vs services)
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18573) InfluxDB reporter cannot be loaded as plugin

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-18573:
-
Summary: InfluxDB reporter cannot be loaded as plugin  (was: 
plugins/metrics-influx/flink-metrics-influxdb-1.11.0.jar META-INF has no dir 
named "services" , but "service")

> InfluxDB reporter cannot be loaded as plugin
> 
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-18573) plugins/metrics-influx/flink-metrics-influxdb-1.11.0.jar META-INF has no dir named "services" , but "service"

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-18573.

Resolution: Fixed

master:
49d3cabb8ae53900138fc1a2f5178dbec46ce378
c3e0d450c82ea8246a7f0abab3a1f86a0f53de7d
1.11:
5c73b0c4a1b50d1b2b3a6a3a3637677a143e598f
60cf377b01595f257992d8d30d1b236d0f6f0103

> plugins/metrics-influx/flink-metrics-influxdb-1.11.0.jar META-INF has no dir 
> named "services" , but "service"
> -
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol merged pull request #12883: [FLINK-18573][metrics] Ensure metric reporters could be loaded via SPI

2020-07-14 Thread GitBox


zentol merged pull request #12883:
URL: https://github.com/apache/flink/pull/12883


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18573) plugins/metrics-influx/flink-metrics-influxdb-1.11.0.jar META-INF has no dir named "services" , but "service"

2020-07-14 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler updated FLINK-18573:
-
Fix Version/s: 1.12.0

> plugins/metrics-influx/flink-metrics-influxdb-1.11.0.jar META-INF has no dir 
> named "services" , but "service"
> -
>
> Key: FLINK-18573
> URL: https://issues.apache.org/jira/browse/FLINK-18573
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics
>Affects Versions: 1.11.0
>Reporter: zhangyunyun
>Assignee: Yun Tang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.1
>
>
> It causes the error:
>  
> 2020-07-13 09:08:46.146 [main] WARN 
> org.apache.flink.runtime.metrics.ReporterSetup - The reporter factory 
> (org.apache.flink.metrics.influxdb.InfluxdbReporterFactory) could not be 
> found for reporter influxdb. Available factories: 
> [org.apache.flink.metrics.slf4j.Slf4jReporterFactory, 
> org.apache.flink.metrics.datadog.DatadogHttpReporterFactory, 
> org.apache.flink.metrics.graphite.GraphiteReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporterFactory, 
> org.apache.flink.metrics.statsd.StatsDReporterFactory, 
> org.apache.flink.metrics.prometheus.PrometheusReporterFactory, 
> org.apache.flink.metrics.jmx.JMXReporterFactory].
> 2020-07-13 09:08:46.149 [main] INFO 
> org.apache.flink.runtime.metrics.MetricRegistryImpl - Periodically reporting 
> metrics in intervals of 60 SECONDS for reporter slf4j of type 
> org.apache.flink.metrics.slf4j.Slf4jReporter.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-11143) AskTimeoutException is thrown during job submission and completion

2020-07-14 Thread Steven Zhen Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-11143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17156363#comment-17156363
 ] 

Steven Zhen Wu edited comment on FLINK-11143 at 7/14/20, 5:38 PM:
--

[~trohrmann] I am seeing a similar problem *when trying unaligned checkpoint 
with 1.11.0*. The Flink job actually started fine. We didn't see this 
AskTimeoutException thrown during job submission in without unaligned 
checkpoint (1.10 or 1.11).

Some more context about the app
 * a large-state stream join app (a few TBs)
 * parallelism 1,440
 * number of containers: 180
 * Cores per container: 12
 * TM_TASK_SLOTS: 8
 * akka.ask.timeout: 120 s
 * heartbeat.timeout: 12
 * web.timeout: 6 (also tried larger values like 300,000 or 600,000 without 
any difference)

I will send you the log files (with DEBUG level) in an email offline. Thanks a 
lot for your help in advance!
{code:java}
\"errors\":[\"Internal server error.\",\" AskTimeoutException is thrown during job submission and completion
> --
>
> Key: FLINK-11143
> URL: https://issues.apache.org/jira/browse/FLINK-11143
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.6.2, 1.10.0
>Reporter: Alex Vinnik
>Priority: Critical
> Attachments: flink-job-timeline.PNG
>
>
> For more details please see the thread
> [http://mail-archives.apache.org/mod_mbox/flink-user/201812.mbox/%3cc2fb26f9-1410-4333-80f4-34807481b...@gmail.com%3E]
> On submission 
> 2018-12-12 02:28:31 ERROR JobsOverviewHandler:92 - Implementation error: 
> Unhandled exception.
>  akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#225683351|#225683351]] after [1 ms]. 
> Sender[null] sent message of type 
> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage".
>  at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604)
>  at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
>  at 
> scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
>  at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236)
>  at java.lang.Thread.run(Thread.java:748)
>  
> On completion
>  
> {"errors":["Internal server error."," side:\njava.util.concurrent.CompletionException: 
> akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#105638574]] after [1 ms]. 
> Sender[null] sent message of type 
> \"org.apache.flink.runtime.rpc.messages.LocalFencedMessage\".
> at 
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:292)
> at 
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:308)
> at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:593)
> at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
> at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
> at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
> at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:772)
> at akka.dispatch.OnComplete.internal(Future.scala:258)
> at akka.dispatch.OnComplete.internal(Future.scala:256)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:186)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:183)
> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:83)
> at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
> at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
> at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:603)
> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
> at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
> at scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
> at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
> at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
> at 
> 

[jira] [Comment Edited] (FLINK-11143) AskTimeoutException is thrown during job submission and completion

2020-07-14 Thread Steven Zhen Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-11143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17156363#comment-17156363
 ] 

Steven Zhen Wu edited comment on FLINK-11143 at 7/14/20, 5:38 PM:
--

[~trohrmann] I am seeing a similar problem *when trying unaligned checkpoint 
with 1.11.0*. The Flink job actually started fine. We didn't see this 
AskTimeoutException thrown during job submission in without unaligned 
checkpoint (1.10 or 1.11).

Some more context about the app
 * a large-state stream join app (a few TBs)
 * parallelism 1,440
 * number of containers: 180
 * Cores per container: 12
 * TM_TASK_SLOTS: 8
 * akka.ask.timeout: 120 s
 * heartbeat.timeout: 12
 * web.timeout: 6 (also tried larger values like 300,000 or 600,000 without 
any difference)

I will send you the log files (with DEBUG level) in an email offline. Thanks a 
lot for your help in advance!
{code:java}
\"errors\":[\"Internal server error.\",\" AskTimeoutException is thrown during job submission and completion
> --
>
> Key: FLINK-11143
> URL: https://issues.apache.org/jira/browse/FLINK-11143
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.6.2, 1.10.0
>Reporter: Alex Vinnik
>Priority: Critical
> Attachments: flink-job-timeline.PNG
>
>
> For more details please see the thread
> [http://mail-archives.apache.org/mod_mbox/flink-user/201812.mbox/%3cc2fb26f9-1410-4333-80f4-34807481b...@gmail.com%3E]
> On submission 
> 2018-12-12 02:28:31 ERROR JobsOverviewHandler:92 - Implementation error: 
> Unhandled exception.
>  akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#225683351|#225683351]] after [1 ms]. 
> Sender[null] sent message of type 
> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage".
>  at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604)
>  at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
>  at 
> scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
>  at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236)
>  at java.lang.Thread.run(Thread.java:748)
>  
> On completion
>  
> {"errors":["Internal server error."," side:\njava.util.concurrent.CompletionException: 
> akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#105638574]] after [1 ms]. 
> Sender[null] sent message of type 
> \"org.apache.flink.runtime.rpc.messages.LocalFencedMessage\".
> at 
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:292)
> at 
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:308)
> at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:593)
> at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
> at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
> at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
> at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:772)
> at akka.dispatch.OnComplete.internal(Future.scala:258)
> at akka.dispatch.OnComplete.internal(Future.scala:256)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:186)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:183)
> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:83)
> at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
> at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
> at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:603)
> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
> at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
> at scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
> at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
> at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
> at 
> 

[jira] [Comment Edited] (FLINK-11143) AskTimeoutException is thrown during job submission and completion

2020-07-14 Thread Steven Zhen Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-11143?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17156363#comment-17156363
 ] 

Steven Zhen Wu edited comment on FLINK-11143 at 7/14/20, 5:37 PM:
--

[~trohrmann] I am seeing a similar problem *when trying unaligned checkpoint 
with 1.11.0*. The Flink job actually started fine. We didn't see this 
AskTimeoutException thrown during job submission in without unaligned 
checkpoint (1.10 or 1.11).

Some more context about the app
 * a large-state stream join app (a few TBs)
 * parallelism 1,440
 * number of containers: 180
 * Cores per container: 12
 * TM_TASK_SLOTS: 8
 * akka.ask.timeout: 120 s
 * heartbeat.timeout: 12
 * web.timeout: 6 (also tried larger values like 300,000 or 600,000 without 
any difference)

I will send you the log files (with DEBUG level) in an email offline. Thanks a 
lot for your help in advance!
{code:java}
\"errors\":[\"Internal server error.\",\" AskTimeoutException is thrown during job submission and completion
> --
>
> Key: FLINK-11143
> URL: https://issues.apache.org/jira/browse/FLINK-11143
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.6.2, 1.10.0
>Reporter: Alex Vinnik
>Priority: Critical
> Attachments: flink-job-timeline.PNG
>
>
> For more details please see the thread
> [http://mail-archives.apache.org/mod_mbox/flink-user/201812.mbox/%3cc2fb26f9-1410-4333-80f4-34807481b...@gmail.com%3E]
> On submission 
> 2018-12-12 02:28:31 ERROR JobsOverviewHandler:92 - Implementation error: 
> Unhandled exception.
>  akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#225683351|#225683351]] after [1 ms]. 
> Sender[null] sent message of type 
> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage".
>  at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604)
>  at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
>  at 
> scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
>  at 
> scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
>  at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284)
>  at 
> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236)
>  at java.lang.Thread.run(Thread.java:748)
>  
> On completion
>  
> {"errors":["Internal server error."," side:\njava.util.concurrent.CompletionException: 
> akka.pattern.AskTimeoutException: Ask timed out on 
> [Actor[akka://flink/user/dispatcher#105638574]] after [1 ms]. 
> Sender[null] sent message of type 
> \"org.apache.flink.runtime.rpc.messages.LocalFencedMessage\".
> at 
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:292)
> at 
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:308)
> at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:593)
> at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
> at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:474)
> at 
> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1977)
> at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:772)
> at akka.dispatch.OnComplete.internal(Future.scala:258)
> at akka.dispatch.OnComplete.internal(Future.scala:256)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:186)
> at akka.dispatch.japi$CallbackBridge.apply(Future.scala:183)
> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
> at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:83)
> at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
> at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
> at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:603)
> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126)
> at 
> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601)
> at scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109)
> at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599)
> at 
> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329)
> at 
> 

[GitHub] [flink] flinkbot edited a comment on pull request #12883: [FLINK-18573][metrics] Ensure metric reporters could be loaded via SPI

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12883:
URL: https://github.com/apache/flink/pull/12883#issuecomment-657495689


   
   ## CI report:
   
   * 25d465768bdf87bd31e0ca9b5749344c35c919d5 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4504)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] twalthr commented on a change in pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


twalthr commented on a change in pull request #12900:
URL: https://github.com/apache/flink/pull/12900#discussion_r454506847



##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/types/DataTypeQueryable.java
##
@@ -0,0 +1,30 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.types;
+
+import org.apache.flink.annotation.PublicEvolving;
+
+/**
+ * Indicates that a {@link DataType} can be retrieved from a class that 
implements this interfaces.
+ */
+@PublicEvolving
+public interface DataTypeQueryable {

Review comment:
   I thought about this as well and decided for keeping `WrapperTypeInfo` 
in the runtime package. The reason for this is to keep the serializers close to 
this utility class in the same package. `WrapperTypeInfo` is not needed in 
`table-common` because it is internal. Actually the constructor of 
`WrappenTypeInfo` should even be private but this is not possible due to other 
failing tests so far. We can mark `DataTypeQueryable` internal if that helps, 
but the past around `TypeInformation` has shown that you need such a simple 
interface over and over again. It will be more useful in the future once we 
introduce a "converter type serializer" where the return data type is not of 
internal data structure.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12898: [FLINK-17636][tests] Fix unstable test SingleInputGateTest#testConcurrentReadStateAndProcessAndClose

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12898:
URL: https://github.com/apache/flink/pull/12898#issuecomment-658161964


   
   ## CI report:
   
   * da4c38d2f9cff86990bef5bf357595971ef1cef9 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4505)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-654978676


   
   ## CI report:
   
   * fbd3e1559f4a80424cbade77ed3034977943bf0b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4463)
 
   * 8c8f594927a2b72c23f8619c1932219585755054 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4509)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-654978676


   
   ## CI report:
   
   * fbd3e1559f4a80424cbade77ed3034977943bf0b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4463)
 
   * 8c8f594927a2b72c23f8619c1932219585755054 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] nielsbasjes edited a comment on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


nielsbasjes edited a comment on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-658267581


   Notes about the current set of changes:
   - Updated the Google dependencies. All Google dependencies no longer specify 
any versions. So in all cases the inclusion of the `libraries-bom` is needed. 
At this point I chose to make it such that downstream projects must specify 
which `libraries-bom` version they need. Although this is an extra step for 
users I hope this will make updating to newer versions of the Google tools 
easier.
   - I fixed almost all cases of Google/Flink api's that have become deprecated 
over time. One clear exception: The use of the ListCheckpointed. This seemed to 
become too large ( a full rewrite of the handling of the acknowledgements ) so 
I think it should be a separate effort.
   - I replaced the use of the NoCredentials and NoCredentialsProvider with 
custom implementations because I ran into problems. See also my bug report 
towards Google https://github.com/googleapis/gax-java/issues/1148
   - I added a full end-to-end test that both reads from and writes to PubSub.
   - I updated the tests to run in parallel 4 (instead of 1) which surfaced a 
small bug that only appears on "ending" streams (like we only have in test 
scenarios). Solution: Move the closing of the connection to a slightly later 
point. This also meant that the `BoundedStringDeserializer` which counted the 
number of messages could no longer be used.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] nielsbasjes commented on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


nielsbasjes commented on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-658270987


   @Xeli @becketqin @rmetzger Since you were involved in reviewing this 
component when originally created I'm tagging you to kindly review these 
changes.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] nielsbasjes commented on pull request #12846: [FLINK-18448][pubsub] Update Google Cloud PubSub dependencies

2020-07-14 Thread GitBox


nielsbasjes commented on pull request #12846:
URL: https://github.com/apache/flink/pull/12846#issuecomment-658267581


   Notes about the current set of changes:
   - Updated the Google dependencies. All Google dependencies no longer specify 
any versions. So in all cases the inclusion of the `libraries-bom` is needed. 
At this point I chose to make it such that downstream projects must specify 
which `libraries-bom` version they need. Although this is an extra step for 
users I hope this will make updating to newer versions of the Google tools 
easier.
   - I fixed almost all cases of Google/Flink api's that have become deprecated 
over time. One clear exception: The use of the ListCheckpointed. This seemed to 
become too large ( a full rewrite of the handling of the acknowledgements ) so 
I think it should be a separate effort.
   - I replaced the use of the NoCredentials and NoCredentialsProvider with 
custom implementations because I ran into problems. See also my bug report 
towards Google https://github.com/googleapis/gax-java/issues/1148
   - I added a full end-to-end test that both reads from and writes to PubSub.
   - I updated the tests to run in parallel 4 (instead of 1) which surfaced a 
small bug that only appears on "ending" streams (like we only have in test 
scenarios). Solution: Move the closing of the connection to a slightly later 
point.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12883: [FLINK-18573][metrics] Ensure metric reporters could be loaded via SPI

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12883:
URL: https://github.com/apache/flink/pull/12883#issuecomment-657495689


   
   ## CI report:
   
   * 93616ee036d921efea80c8003e408d35339cc4ec Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4500)
 
   * 25d465768bdf87bd31e0ca9b5749344c35c919d5 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4504)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] dawidwys commented on a change in pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


dawidwys commented on a change in pull request #12900:
URL: https://github.com/apache/flink/pull/12900#discussion_r454415036



##
File path: 
flink-table/flink-table-common/src/main/java/org/apache/flink/table/types/DataTypeQueryable.java
##
@@ -0,0 +1,30 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.types;
+
+import org.apache.flink.annotation.PublicEvolving;
+
+/**
+ * Indicates that a {@link DataType} can be retrieved from a class that 
implements this interfaces.
+ */
+@PublicEvolving
+public interface DataTypeQueryable {

Review comment:
   Is this interface really necessary? Especially with `@PublicEvolving` 
annotation? How are users supposed to use it? If I understand it correctly you 
need it for internal operations. Moreover you need it because the 
`WrapperTypeInfo` is in `blink-runtime`, right?
   
   Can't we move the `WrapperTypeInfo` to the `table-common` instead?  The 
class itself has no runtime dependencies. Only the factory methods need some 
runtime classes.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12900:
URL: https://github.com/apache/flink/pull/12900#issuecomment-658213136


   
   ## CI report:
   
   * ceb895be265bd445b9d36bfc52cf8018a508c333 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4508)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18588) hive ddl create table should support 'if not exists'

2020-07-14 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157430#comment-17157430
 ] 

Jark Wu commented on FLINK-18588:
-

[~wtog] I CC'ed someone familiar with this part to review in the PR. 

> hive ddl create table should support 'if not exists'
> 
>
> Key: FLINK-18588
> URL: https://issues.apache.org/jira/browse/FLINK-18588
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.11.0
>Reporter: wangtong
>Priority: Minor
>  Labels: pull-request-available
>
> FlinkHiveSqlParser cant parse sql 'create table if not exists tbl (x int)' 
> due to 'if not exists'
> {code:java}
> // code placeholder
> java.lang.RuntimeException: Error while parsing SQL: create table if not 
> exists tbl (x int)java.lang.RuntimeException: Error while parsing SQL: create 
> table if not exists tbl (x int)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$TesterImpl.parseStmtAndHandleEx(SqlParserTest.java:8861)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$TesterImpl.check(SqlParserTest.java:8845)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$Sql.ok(SqlParserTest.java:9192) 
> at 
> org.apache.flink.sql.parser.hive.FlinkHiveSqlParserImplTest.testCreateTable(FlinkHiveSqlParserImplTest.java:177)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>  at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>  at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363) at 
> org.junit.runner.JUnitCore.run(JUnitCore.java:137) at 
> org.junit.runner.JUnitCore.run(JUnitCore.java:115) at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:40)
>  at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:184) 
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) 
> at java.util.Iterator.forEachRemaining(Iterator.java:116) at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>  at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:482) at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:472) 
> at 
> java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:151) 
> at 
> java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:174)
>  at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at 
> java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:418) at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>  at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:71) 
> at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:229)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$6(DefaultLauncher.java:197)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:211)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:191)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:128)
>  at 
> com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:69)
>  at 
> com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
>  at 
> com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:230)
>  at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:58)Caused by: 
> org.apache.calcite.sql.parser.SqlParseException: Encountered "not" at line 1, 
> column 17.Was expecting one of:         

[GitHub] [flink] wuchong commented on pull request #12888: [FLINK-18588] hive ddl create table support 'if not exists'

2020-07-14 Thread GitBox


wuchong commented on pull request #12888:
URL: https://github.com/apache/flink/pull/12888#issuecomment-658219194


   cc @lirui-apache @JingsongLi 
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


flinkbot commented on pull request #12900:
URL: https://github.com/apache/flink/pull/12900#issuecomment-658213136


   
   ## CI report:
   
   * ceb895be265bd445b9d36bfc52cf8018a508c333 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12899: [FLINK-16699][k8s] Support accessing secured services via K8s secrets

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12899:
URL: https://github.com/apache/flink/pull/12899#issuecomment-658162074


   
   ## CI report:
   
   * 77f95464c6036283d77d6e81e2bd21ab27f29d54 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4506)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


flinkbot commented on pull request #12900:
URL: https://github.com/apache/flink/pull/12900#issuecomment-658202311


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit ceb895be265bd445b9d36bfc52cf8018a508c333 (Tue Jul 14 
14:09:33 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] twalthr commented on pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


twalthr commented on pull request #12900:
URL: https://github.com/apache/flink/pull/12900#issuecomment-658201596


   CC @wuchong @dawidwys
   Btw after all this refactoring I was thinking about renaming 
`WrapperTypeInfo` to `InternalTypeInfo`. This would better match to its 
purpose. What do you think?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] twalthr opened a new pull request #12900: [FLINK-18537][table] Remove RowDataTypeInfo

2020-07-14 Thread GitBox


twalthr opened a new pull request #12900:
URL: https://github.com/apache/flink/pull/12900


   ## What is the purpose of the change
   
   `TypeInformation` is a legacy class for the sole purpose of creating a 
`TypeSerializer`.
Instances of `TypeInformation` are not required in the table ecosystem but 
sometimes enforced
by interfaces of other modules (such as 
`org.apache.flink.api.dag.Transformation`). Therefore, we recently introduced 
`WrapperTypeInfo` which acts as an adapter whenever type information is 
required. Instances of `WrapperTypeInfo` should only be created for passing it 
to interfaces that require type information. The class should not be used as a 
replacement for a `LogicalType`. Information such as the arity of a row type, 
field types, field names, etc. should be derived from the `LogicalType` 
directly.
   
   This PR updates the code base to not use `RowDataTypeInfo` anymore. 
`RowDataTypeInfo` was a composite type information with legacy types as field 
types. This refactoring unblocks FLINK-15803 and reduces potential bugs from 
type conversion. 
   
   ## Brief change log
   
   See commit messages.
   
   ## Verifying this change
   
   This change is already covered by existing tests.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: yes
 - The serializers: yes
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18537) Remove RowDataTypeInfo

2020-07-14 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18537?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18537:
---
Labels: pull-request-available  (was: )

> Remove RowDataTypeInfo
> --
>
> Key: FLINK-18537
> URL: https://issues.apache.org/jira/browse/FLINK-18537
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Timo Walther
>Assignee: Timo Walther
>Priority: Major
>  Labels: pull-request-available
>
> FLINK-17000 introduced a TypeInformation class that should replace most of 
> the type information in the Blink planner. We start with removing 
> RowDataTypeInfo.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18588) hive ddl create table should support 'if not exists'

2020-07-14 Thread wangtong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17157401#comment-17157401
 ] 

wangtong commented on FLINK-18588:
--

[~jark] please help to review this pr

> hive ddl create table should support 'if not exists'
> 
>
> Key: FLINK-18588
> URL: https://issues.apache.org/jira/browse/FLINK-18588
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.11.0
>Reporter: wangtong
>Priority: Minor
>  Labels: pull-request-available
>
> FlinkHiveSqlParser cant parse sql 'create table if not exists tbl (x int)' 
> due to 'if not exists'
> {code:java}
> // code placeholder
> java.lang.RuntimeException: Error while parsing SQL: create table if not 
> exists tbl (x int)java.lang.RuntimeException: Error while parsing SQL: create 
> table if not exists tbl (x int)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$TesterImpl.parseStmtAndHandleEx(SqlParserTest.java:8861)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$TesterImpl.check(SqlParserTest.java:8845)
>  at 
> org.apache.calcite.sql.parser.SqlParserTest$Sql.ok(SqlParserTest.java:9192) 
> at 
> org.apache.flink.sql.parser.hive.FlinkHiveSqlParserImplTest.testCreateTable(FlinkHiveSqlParserImplTest.java:177)
>  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
> at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>  at java.lang.reflect.Method.invoke(Method.java:498) at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>  at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>  at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363) at 
> org.junit.runner.JUnitCore.run(JUnitCore.java:137) at 
> org.junit.runner.JUnitCore.run(JUnitCore.java:115) at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:40)
>  at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:184) 
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) 
> at java.util.Iterator.forEachRemaining(Iterator.java:116) at 
> java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
>  at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:482) at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:472) 
> at 
> java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:151) 
> at 
> java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:174)
>  at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at 
> java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:418) at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>  at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:71) 
> at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:229)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.lambda$execute$6(DefaultLauncher.java:197)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.withInterceptedStreams(DefaultLauncher.java:211)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:191)
>  at 
> org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:128)
>  at 
> com.intellij.junit5.JUnit5IdeaTestRunner.startRunnerWithArgs(JUnit5IdeaTestRunner.java:69)
>  at 
> com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
>  at 
> com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:230)
>  at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:58)Caused by: 
> org.apache.calcite.sql.parser.SqlParseException: Encountered "not" at line 1, 
> column 17.Was expecting one of:         "ROW" ...    "COMMENT" ...    
> 

[GitHub] [flink] flinkbot edited a comment on pull request #12866: [FLINK-17425][blink-planner] supportsFilterPushDown rule in DynamicSource.

2020-07-14 Thread GitBox


flinkbot edited a comment on pull request #12866:
URL: https://github.com/apache/flink/pull/12866#issuecomment-656576219


   
   ## CI report:
   
   * 02c81cff07ef76dc20a7c618112cc7a172856720 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=4497)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




  1   2   3   >