[GitHub] [flink] flinkbot edited a comment on pull request #12283: [FLINK-16975][documentation] Add docs for FileSystem connector

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12283:
URL: https://github.com/apache/flink/pull/12283#issuecomment-632089857


   
   ## CI report:
   
   * b3fd51b309c78d9dd5056eed35dc2fe388665899 UNKNOWN
   * c29ec64e9bc260015c4b57395ace4a578926fa16 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2383)
 
   * 8569d299f41b57033debc94d33ab11af91452140 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2689)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zentol commented on a change in pull request #12447: [FLINK-18069][CI] Test if Java/Scaladocs builds are passing in the compile stage

2020-06-04 Thread GitBox


zentol commented on a change in pull request #12447:
URL: https://github.com/apache/flink/pull/12447#discussion_r435040079



##
File path: tools/ci/compile.sh
##
@@ -54,10 +54,22 @@ if [ $EXIT_CODE != 0 ]; then
 exit $EXIT_CODE
 fi
 
-echo " Checking Scaladocs "
+echo " Checking Javadocs and Scaladocs "
+
+# use the same invocation as on buildbot 
(https://svn.apache.org/repos/infra/infrastructure/buildbot/aegis/buildmaster/master1/projects/flink.conf)
+run_mvn javadoc:aggregate -Paggregate-scaladoc 
-DadditionalJOption='-Xdoclint:none' \

Review comment:
   same deal as scala, move the output into a file.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] dawidwys closed pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


dawidwys closed pull request #12455:
URL: https://github.com/apache/flink/pull/12455


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #11541: [FLINK-15416][network] add task manager netty client retry mechenism

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #11541:
URL: https://github.com/apache/flink/pull/11541#issuecomment-604812212


   
   ## CI report:
   
   * b5ef19285000d692d6b1b8b0aa7a37ce35c216f5 UNKNOWN
   * baf2c55f7bb0b9df7bdef94083c947693b75eae2 UNKNOWN
   * e5419f41a18eaae05814837d679f58480a2cb94c UNKNOWN
   * a5dc1df6976661f4dc56c5e739b54dd90efc660d Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/169527021) Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2679)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12474: [javadocs] Fix typo in TwoPhaseCommitSinkFunction.java

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12474:
URL: https://github.com/apache/flink/pull/12474#issuecomment-638579512


   
   ## CI report:
   
   * 7c4289c4f4b7752e5693f631c4e84c1ebc0d8beb Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2675)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12410: [FLINK-13782][table-api] Implement type strategies for IF ELSE expression

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12410:
URL: https://github.com/apache/flink/pull/12410#issuecomment-635946149


   
   ## CI report:
   
   * ce34ce1e7b16cf8c6bf359ffadbf5d3e96613947 UNKNOWN
   * 9e2c9729f2c4fb3af0e339c9f7046bf1cef2fb60 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2659)
 
   * 5e2c5e2f0d9babcc26b6223449b3ab6f8ac97c3e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] carp84 commented on a change in pull request #12144: [FLINK-17384][flink-dist] support read hbase conf dir from flink.conf and change HBaseConfiguration construction.

2020-06-04 Thread GitBox


carp84 commented on a change in pull request #12144:
URL: https://github.com/apache/flink/pull/12144#discussion_r435046468



##
File path: 
flink-connectors/flink-connector-hbase/src/main/java/org/apache/flink/connector/hbase/util/HBaseConfigurationUtil.java
##
@@ -22,20 +22,85 @@
 import org.apache.flink.util.Preconditions;
 
 import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.hbase.HBaseConfiguration;
 import org.apache.hadoop.io.Writable;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 import java.io.ByteArrayInputStream;
 import java.io.ByteArrayOutputStream;
 import java.io.DataInputStream;
 import java.io.DataOutputStream;
+import java.io.File;
 import java.io.IOException;
 
 /**
- * This class helps to do serialization for hadoop Configuration.
+ * This class helps to do serialization for hadoop Configuration and 
HBase-related classes.
  */
 @Internal
 public class HBaseConfigurationUtil {
 
+   private static final Logger LOG = 
LoggerFactory.getLogger(HBaseConfigurationUtil.class);
+
+   @SuppressWarnings("deprecation")
+   public static Configuration getHBaseConfiguration() {
+
+   // Instantiate an HBaseConfiguration to load the 
hbase-default.xml and hbase-site.xml from the classpath.
+   Configuration result = new HBaseConfiguration().create();

Review comment:
   It seems this line is not updated in the new commit, maybe you missed 
the change? @liuyongvs  Thanks.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong commented on a change in pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


wuchong commented on a change in pull request #12436:
URL: https://github.com/apache/flink/pull/12436#discussion_r435038062



##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1648,25 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+index.literalValue match {
+  case Some(v: Int) if v < 1 =>
+throw new ValidationException(
+  s"Array element access needs an index starting at 1 but was $v.")
+  case _ => //nothing
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
 val arrayAccessCode =
-  s"""
- |${array.code}
- |${index.code}
- |boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull;
- |$resultTypeTerm $resultTerm = $nullTerm ? $defaultTerm : $arrayGet;
- |""".stripMargin
-
+s"""
+|${array.code}
+|${index.code}
+|boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull ||

Review comment:
   Move `$arrayIsNull` to the end? Otherwise, it is possible to throw 
`ArrayIndexOutOfBoundsException`. You can try to test `f11[3]`. Maybe we should 
rename `arrayIsNull` to `arrayElementIsNull`.

##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1648,25 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+index.literalValue match {
+  case Some(v: Int) if v < 1 =>
+throw new ValidationException(
+  s"Array element access needs an index starting at 1 but was $v.")
+  case _ => //nothing
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
 val arrayAccessCode =
-  s"""
- |${array.code}
- |${index.code}
- |boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull;
- |$resultTypeTerm $resultTerm = $nullTerm ? $defaultTerm : $arrayGet;
- |""".stripMargin
-
+s"""
+|${array.code}
+|${index.code}
+|boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull ||
+|   ${idxStr} < 0 || ${idxStr} >= ${array.resultTerm}.size();

Review comment:
   `${idxStr}` -> `$idxStr`.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-15066) Cannot run multiple `insert into csvTable values ()`

2020-06-04 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski updated FLINK-15066:
---
Affects Version/s: 1.10.1

> Cannot run multiple `insert into csvTable values ()`
> 
>
> Key: FLINK-15066
> URL: https://issues.apache.org/jira/browse/FLINK-15066
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.10.1
>Reporter: Kurt Young
>Assignee: Jingsong Lee
>Priority: Blocker
> Fix For: 1.11.0
>
>
> I created a csv table in sql client, and tried to insert some data into this 
> table.
> The first insert into success, but the second one failed with exception: 
> {code:java}
> // Caused by: java.io.IOException: File or directory /.../xxx.csv already 
> exists. Existing files and directories are not overwritten in NO_OVERWRITE 
> mode. Use OVERWRITE mode to overwrite existing files and directories.at 
> org.apache.flink.core.fs.FileSystem.initOutPathLocalFS(FileSystem.java:817)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] lsyldliu commented on pull request #12303: [FLINK-17625] [table] Fix ArrayIndexOutOfBoundsException in AppendOnlyTopNFunction

2020-06-04 Thread GitBox


lsyldliu commented on pull request #12303:
URL: https://github.com/apache/flink/pull/12303#issuecomment-638737701


   cc @beyond1920



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] lsyldliu commented on pull request #12303: [FLINK-17625] [table] Fix ArrayIndexOutOfBoundsException in AppendOnlyTopNFunction

2020-06-04 Thread GitBox


lsyldliu commented on pull request #12303:
URL: https://github.com/apache/flink/pull/12303#issuecomment-638736205


   cc @wuchong 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Comment Edited] (FLINK-18109) Manually test external resource framework with GPUDriver

2020-06-04 Thread Yangze Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125740#comment-17125740
 ] 

Yangze Guo edited comment on FLINK-18109 at 6/4/20, 9:40 AM:
-

Test log:
 * Standalone mode, Linux, 2 NVIDIA GPUs
 ** normal mode, works well.
 ** coordination mode, works well.
 * Kubernetes mode
 ** build a custom image from Flink-1.11.0-RC1
 ** {{bin/kubernetes-session.sh and submit WordCount job}}
 ** log of task executor shows everything works well.
 * Yarn
 ** 2.10, works well.
 ** 3.1, works well.


was (Author: karmagyz):
Test log:
 * Standalone mode, Linux, 2 NVIDIA GPUs
 ** normal mode, works well.
 ** coordination mode, works well.
 * Kubernetes mode
 ** build a custom image from Flink-1.11.0-RC1
 ** {{bin/kubernetes-session.sh and submit WordCount job}}
 ** log of task executor shows everything works well
 * Yarn
 ** Todo

> Manually test external resource framework with GPUDriver
> 
>
> Key: FLINK-18109
> URL: https://issues.apache.org/jira/browse/FLINK-18109
> Project: Flink
>  Issue Type: Sub-task
>Affects Versions: 1.11.0
>Reporter: Yangze Guo
>Priority: Blocker
> Fix For: 1.11.0
>
>
> We need to:
>  - Test it in standalone mode.
>  ** The discovery script works well in default mode
>  ** The discovery script works well in coordination mode.
>  - Test it with Yarn 2.10 and 3.1.
>  ** GPU resources are allocated successfully.
>  - Test it with Kubernetes
>  ** GPU resources are allocated successfully.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] JingsongLi commented on a change in pull request #12319: [FLINK-17406][doc] Add documentation about dynamic table options

2020-06-04 Thread GitBox


JingsongLi commented on a change in pull request #12319:
URL: https://github.com/apache/flink/pull/12319#discussion_r435129153



##
File path: docs/dev/table/sql/hints.md
##
@@ -0,0 +1,73 @@
+---
+title: "SQL Hints"
+nav-parent_id: sql
+nav-pos: 6
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+

Review comment:
   Add one line to explain Table hints. This is a doc for hints, not only 
for `Dynamic Table Options`





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] JingsongLi commented on a change in pull request #12319: [FLINK-17406][doc] Add documentation about dynamic table options

2020-06-04 Thread GitBox


JingsongLi commented on a change in pull request #12319:
URL: https://github.com/apache/flink/pull/12319#discussion_r435129153



##
File path: docs/dev/table/sql/hints.md
##
@@ -0,0 +1,73 @@
+---
+title: "SQL Hints"
+nav-parent_id: sql
+nav-pos: 6
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+

Review comment:
   Add one line to explain Table hints.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong commented on pull request #12381: [FLINK-17941][sql-client] Switching catalog or database doesn't work …

2020-06-04 Thread GitBox


wuchong commented on pull request #12381:
URL: https://github.com/apache/flink/pull/12381#issuecomment-638742371


   This has been addressed by #12431.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12144: [FLINK-17384][flink-dist] support read hbase conf dir from flink.conf and change HBaseConfiguration construction.

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12144:
URL: https://github.com/apache/flink/pull/12144#issuecomment-628475376


   
   ## CI report:
   
   * d36b959fb16a91c15babda10dd884cbbdec58420 UNKNOWN
   * 33b1ef0748285bc503363458a684ab237a64cd62 UNKNOWN
   * e701eb7677b535e07e4d5642ae9280cab084b854 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2668)
 
   * 30ac550d12c2eca22eaa4dab8f9186d7296355cd Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2703)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18055) Catalog does not exist in SQL Client

2020-06-04 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18055?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125756#comment-17125756
 ] 

Jark Wu commented on FLINK-18055:
-

[~dwysakowicz], I discussed with [~lirui] and yes, FLINK-17941 is a duplicate. 
We both agree to use the pull request of this issue. 

> Catalog does not exist in SQL Client
> 
>
> Key: FLINK-18055
> URL: https://issues.apache.org/jira/browse/FLINK-18055
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.11.0
>Reporter: godfrey he
>Assignee: godfrey he
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Flink SQL> show catalogs;
> default_catalog
> hive
> Flink SQL> use  catalog hive;
> [ERROR] Could not execute SQL statement. Reason:
> org.apache.flink.table.catalog.exceptions.CatalogException: A catalog with 
> name [`hive`] does not exist.
> The reason is {{SqlCommandParser}} adds {{``}} for catalog name, which is 
> unnecessary. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-18055) Catalog does not exist in SQL Client

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18055?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-18055.
---
Resolution: Fixed

- master (1.12.0): aa40ffd94a19ed9b57988d42b8c9262ef60e172e
- 1.11.0: f391b0d433ed16f33018837f42c383992953a7bc

> Catalog does not exist in SQL Client
> 
>
> Key: FLINK-18055
> URL: https://issues.apache.org/jira/browse/FLINK-18055
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.11.0
>Reporter: godfrey he
>Assignee: godfrey he
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Flink SQL> show catalogs;
> default_catalog
> hive
> Flink SQL> use  catalog hive;
> [ERROR] Could not execute SQL statement. Reason:
> org.apache.flink.table.catalog.exceptions.CatalogException: A catalog with 
> name [`hive`] does not exist.
> The reason is {{SqlCommandParser}} adds {{``}} for catalog name, which is 
> unnecessary. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-17941) Switching catalog or database doesn't work from SQL CLI

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17941?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-17941.
---
Resolution: Duplicate

> Switching catalog or database doesn't work from SQL CLI
> ---
>
> Key: FLINK-17941
> URL: https://issues.apache.org/jira/browse/FLINK-17941
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Reporter: Rui Li
>Assignee: Rui Li
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> To reproduce:
> {noformat}
> Flink SQL> create database db1;
> [INFO] Database has been created.
> Flink SQL> use db1;
> [ERROR] Could not execute SQL statement. Reason:
> org.apache.flink.table.catalog.exceptions.CatalogException: A database with 
> name [`myhive`.`db1`] does not exist in the catalog: [myhive].
> Flink SQL> show databases;
> db1
> default
> {noformat}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-docker] KeatsPeeks commented on pull request #9: [FLINK-16260][docker] Add docker images based on Java 11

2020-06-04 Thread GitBox


KeatsPeeks commented on pull request #9:
URL: https://github.com/apache/flink-docker/pull/9#issuecomment-638750132


   Any update on this one ?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18121) Support creating Docker image from local Flink distribution

2020-06-04 Thread Till Rohrmann (Jira)
Till Rohrmann created FLINK-18121:
-

 Summary: Support creating Docker image from local Flink 
distribution
 Key: FLINK-18121
 URL: https://issues.apache.org/jira/browse/FLINK-18121
 Project: Flink
  Issue Type: Improvement
  Components: Dockerfiles
Affects Versions: docker-1.11.0.0
Reporter: Till Rohrmann


Currently, 
https://github.com/apache/flink-docker/blob/dev-master/Dockerfile-debian.template
 only supports to create a Docker image from a Flink distribution which is 
hosted on a web server. I think it would be helpful if we could also create a 
Docker image from a Flink distribution which is stored on one's local file 
system. That way, one would not have to upload the file or start a web server 
for serving it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zentol commented on pull request #12479: [FLINK-17260] Fix StreamingKafkaITCase instabilities

2020-06-04 Thread GitBox


zentol commented on pull request #12479:
URL: https://github.com/apache/flink/pull/12479#issuecomment-638753608


   We should do the same for the 
`org.apache.flink.tests.util.kafka.SQLClientKafkaITCase`



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18121) Support creating Docker image from local Flink distribution

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125783#comment-17125783
 ] 

Robert Metzger commented on FLINK-18121:


FYI: I opened a PR documenting the current approach: 
https://github.com/apache/flink-docker/pull/24

> Support creating Docker image from local Flink distribution
> ---
>
> Key: FLINK-18121
> URL: https://issues.apache.org/jira/browse/FLINK-18121
> Project: Flink
>  Issue Type: Improvement
>  Components: Dockerfiles
>Affects Versions: docker-1.11.0.0
>Reporter: Till Rohrmann
>Priority: Major
>
> Currently, 
> https://github.com/apache/flink-docker/blob/dev-master/Dockerfile-debian.template
>  only supports to create a Docker image from a Flink distribution which is 
> hosted on a web server. I think it would be helpful if we could also create a 
> Docker image from a Flink distribution which is stored on one's local file 
> system. That way, one would not have to upload the file or start a web server 
> for serving it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12460: [FLINK-18063][checkpointing] Fix the race condition for aborting current checkpoint in CheckpointBarrierUnaligner

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12460:
URL: https://github.com/apache/flink/pull/12460#issuecomment-638102290


   
   ## CI report:
   
   * 1d838c03d2b9f9744cae6fc03a919db72cc0efd9 UNKNOWN
   * b2c4d6aaf78612e7214cc49d22e2b90fe29b93b6 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2686)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12455:
URL: https://github.com/apache/flink/pull/12455#issuecomment-638007072


   
   ## CI report:
   
   * 7e9143f4531e9a7477a1e2848020d7c69644a3be Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2666)
 
   * 278da87279e640156951deef07ee9e4e24cab361 UNKNOWN
   * 76cf3b92819b6a77bf9c50b6d7674f5df3f4c8f7 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12352: [FLINK-17717][sql-parser] Throws for DDL create temporary system func…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12352:
URL: https://github.com/apache/flink/pull/12352#issuecomment-634517986


   
   ## CI report:
   
   * 6900843764e0fe637babcb429194ef620be73b50 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2278)
 
   * 7dc817125d0684987814d45192eab87a41ebc5b5 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-637399261


   
   ## CI report:
   
   * d553feca300e2668e3e5c16c1c524d73acde141c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2627)
 
   * 4d1f156fd758a44f4c394a478b6f37bc2de61c86 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2711)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12456: [FLINK-17113][sql-cli] Refactor view support in SQL Client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12456:
URL: https://github.com/apache/flink/pull/12456#issuecomment-638032908


   
   ## CI report:
   
   * b5b8172d4ccddf197a084d38e6dcef13ee27da8e Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2614)
 
   * 12e3932ed8d74a59c0d4c28441b336284a2dffeb Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2713)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12481: [FLINK-18120] Don't expand documentation sections by default

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12481:
URL: https://github.com/apache/flink/pull/12481#issuecomment-638746420


   
   ## CI report:
   
   * 57276170065599217eaf7a77f451d74a9311a870 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2714)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong closed pull request #12441: [FLINK-17918][table-blink] Fix AppendOnlyTopNFunction shouldn't mutate list value of MapState

2020-06-04 Thread GitBox


wuchong closed pull request #12441:
URL: https://github.com/apache/flink/pull/12441


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong commented on pull request #12441: [FLINK-17918][table-blink] Fix AppendOnlyTopNFunction shouldn't mutate list value of MapState

2020-06-04 Thread GitBox


wuchong commented on pull request #12441:
URL: https://github.com/apache/flink/pull/12441#issuecomment-638761387


   Thanks for the reviewing @AHeise . Will merge this. 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] wuchong commented on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


wuchong commented on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-638765700


   Do you have other concerns @libenchao ?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12435: [FLINK-18059] [sql-client] Fix create/drop catalog statement can not be executed in sql client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12435:
URL: https://github.com/apache/flink/pull/12435#issuecomment-637372360


   
   ## CI report:
   
   * b5ef005c90fb95a63c97d06c488adf5e4abe2972 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2626)
 
   * 223701abe70c0815eb1f35af7383727385ce2e90 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-637399261


   
   ## CI report:
   
   * d553feca300e2668e3e5c16c1c524d73acde141c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2627)
 
   * 4d1f156fd758a44f4c394a478b6f37bc2de61c86 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2711)
 
   * 63831f0ba3dc2a6ec31acc7fb1cfd1b51288945b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12326: [FLINK-16577] [table-planner-blink] Fix numeric type mismatch error in  column interval relmetadata

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12326:
URL: https://github.com/apache/flink/pull/12326#issuecomment-633786013


   
   ## CI report:
   
   * 0b14ce7cf01c6664a3a98c35dd1c494a7ad4fc32 UNKNOWN
   * 23311a70dcb1522127aaadf906875d536b64c166 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2171)
 
   * 8c198bc1bb0b6be3321a01898e09d7492169b47e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18071) CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-18071:
---
Affects Version/s: 1.11.0

> CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI
> 
>
> Key: FLINK-18071
> URL: https://issues.apache.org/jira/browse/FLINK-18071
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> CI: 
> https://dev.azure.com/georgeryan1322/Flink/_build/results?buildId=330=logs=6e58d712-c5cc-52fb-0895-6ff7bd56c46b=f30a8e80-b2cf-535c-9952-7f521a4ae374
> {code}
> [ERROR] Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 8.795 
> s <<< FAILURE! - in 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase
> [ERROR] 
> test(org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase)
>   Time elapsed: 4.647 s  <<< FAILURE!
> java.lang.AssertionError: List did not contain expected sequence of 200 
> elements, but was: [152, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 
> 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 
> 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 
> 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 
> 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 
> 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 
> 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 
> 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 
> 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 
> 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 
> 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 
> 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 
> 198, 199]
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.failList(CoordinatorEventsExactlyOnceITCase.java:160)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.checkListContainsSequence(CoordinatorEventsExactlyOnceITCase.java:148)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.test(CoordinatorEventsExactlyOnceITCase.java:143)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18122) Kubernetes test fails with "error: timed out waiting for the condition on jobs/flink-job-cluster"

2020-06-04 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18122:
--

 Summary: Kubernetes test fails with "error: timed out waiting for 
the condition on jobs/flink-job-cluster"
 Key: FLINK-18122
 URL: https://issues.apache.org/jira/browse/FLINK-18122
 Project: Flink
  Issue Type: Bug
  Components: Deployment / Kubernetes, Tests
Affects Versions: 1.11.0
Reporter: Robert Metzger


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2697=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5

{code}
2020-06-04T09:25:40.7205843Z service/flink-job-cluster created
2020-06-04T09:25:40.9661515Z job.batch/flink-job-cluster created
2020-06-04T09:25:41.2189123Z deployment.apps/flink-task-manager created
2020-06-04T10:32:32.6402983Z error: timed out waiting for the condition on 
jobs/flink-job-cluster
2020-06-04T10:32:33.8057757Z error: unable to upgrade connection: container not 
found ("flink-task-manager")
2020-06-04T10:32:33.8111302Z sort: cannot read: 
'/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*':
 No such file or directory
2020-06-04T10:32:33.8124455Z FAIL WordCount: Output hash mismatch.  Got 
d41d8cd98f00b204e9800998ecf8427e, expected e682ec6622b5e83f2eb614617d5ab2cf.
2020-06-04T10:32:33.8125379Z head hexdump of actual:
2020-06-04T10:32:33.8136133Z head: cannot open 
'/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*'
 for reading: No such file or directory
2020-06-04T10:32:33.8344715Z Debugging failed Kubernetes test:
2020-06-04T10:32:33.8345469Z Currently existing Kubernetes resources
2020-06-04T10:32:36.4977853Z I0604 10:32:36.497383   13191 request.go:621] 
Throttling request took 1.198606989s, request: 
GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
2020-06-04T10:32:46.6975735Z I0604 10:32:46.697234   13191 request.go:621] 
Throttling request took 4.398107353s, request: 
GET:https://10.1.0.4:8443/apis/authorization.k8s.io/v1?timeout=32s
2020-06-04T10:32:57.4978637Z I0604 10:32:57.497209   13191 request.go:621] 
Throttling request took 1.198449167s, request: 
GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
2020-06-04T10:33:07.4980104Z I0604 10:33:07.497320   13191 request.go:621] 
Throttling request took 4.198274438s, request: 
GET:https://10.1.0.4:8443/apis/apiextensions.k8s.io/v1?timeout=32s
2020-06-04T10:33:18.4976060Z I0604 10:33:18.497258   13191 request.go:621] 
Throttling request took 1.19871495s, request: 
GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
2020-06-04T10:33:28.4979129Z I0604 10:33:28.497276   13191 request.go:621] 
Throttling request took 4.198369672s, request: 
GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
2020-06-04T10:33:30.9182069Z NAME READY   
STATUS  RESTARTS   AGE
2020-06-04T10:33:30.9184099Z pod/flink-job-cluster-dtb67  0/1 
ErrImageNeverPull   0  67m
2020-06-04T10:33:30.9184869Z pod/flink-task-manager-74ccc9bd9-psqwm   0/1 
ErrImageNeverPull   0  67m
2020-06-04T10:33:30.9185226Z 
2020-06-04T10:33:30.9185926Z NAMETYPECLUSTER-IP 
 EXTERNAL-IP   PORT(S)  
 AGE
2020-06-04T10:33:30.9186832Z service/flink-job-cluster   NodePort
10.111.92.199   
6123:32501/TCP,6124:31360/TCP,6125:30025/TCP,8081:30081/TCP   67m
2020-06-04T10:33:30.9187545Z service/kubernetes  ClusterIP   10.96.0.1  
 443/TCP  
 68m
2020-06-04T10:33:30.9187976Z 
2020-06-04T10:33:30.9188472Z NAME READY   
UP-TO-DATE   AVAILABLE   AGE
2020-06-04T10:33:30.9189179Z deployment.apps/flink-task-manager   0/1 1 
   0   67m
2020-06-04T10:33:30.9189508Z 
2020-06-04T10:33:30.9189815Z NAME   
DESIRED   CURRENT   READY   AGE
2020-06-04T10:33:30.9190418Z replicaset.apps/flink-task-manager-74ccc9bd9   1   
  1 0   67m
2020-06-04T10:33:30.9190662Z 
2020-06-04T10:33:30.9190891Z NAME  COMPLETIONS   
DURATION   AGE
2020-06-04T10:33:30.9191423Z job.batch/flink-job-cluster   0/1   67m
67m
2020-06-04T10:33:33.7840921Z I0604 10:33:33.783675   13482 request.go:621] 
Throttling request took 1.198522435s, request: 
GET:https://10.1.0.4:8443/apis/batch/v1beta1?timeout=32s
2020-06-04T10:33:43.7842354Z I0604 10:33:43.783896   13482 request.go:621] 
Throttling request took 4.198512626s, request: 
GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
2020-06-04T10:33:54.7840871Z I0604 10:33:54.783716   13482 request.go:621] 
Throttling request took 1.198557318s, request: 

[jira] [Updated] (FLINK-18071) CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18071?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-18071:
---
Fix Version/s: 1.11.0

> CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI
> 
>
> Key: FLINK-18071
> URL: https://issues.apache.org/jira/browse/FLINK-18071
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.11.0, 1.12.0
>
>
> CI: 
> https://dev.azure.com/georgeryan1322/Flink/_build/results?buildId=330=logs=6e58d712-c5cc-52fb-0895-6ff7bd56c46b=f30a8e80-b2cf-535c-9952-7f521a4ae374
> {code}
> [ERROR] Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 8.795 
> s <<< FAILURE! - in 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase
> [ERROR] 
> test(org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase)
>   Time elapsed: 4.647 s  <<< FAILURE!
> java.lang.AssertionError: List did not contain expected sequence of 200 
> elements, but was: [152, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 
> 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 
> 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 
> 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 
> 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 
> 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 
> 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 
> 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 
> 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 
> 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 
> 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 
> 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 
> 198, 199]
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.failList(CoordinatorEventsExactlyOnceITCase.java:160)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.checkListContainsSequence(CoordinatorEventsExactlyOnceITCase.java:148)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.test(CoordinatorEventsExactlyOnceITCase.java:143)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] pnowojski edited a comment on pull request #11541: [FLINK-15416][network] add task manager netty client retry mechenism

2020-06-04 Thread GitBox


pnowojski edited a comment on pull request #11541:
URL: https://github.com/apache/flink/pull/11541#issuecomment-638644705


   There are a couple of things left to do there. 
   
   1. For example test is marked with Ignore and I’m not sure if that’s a 
mistake, as the commit marking it ignored seemed to be fixing something. But it 
was 5 years ago so who knows… removing Ignore might introduce test instability, 
so it would have to be thoroughly checked.
   2. There are rebase conflicts.
   3. I think in your test you should be joining the spawned threads and 
checking their’s exceptions.
   
   If you prefer we can take over those issues and merge this PR in a couple of 
weeks - after release 1.11 testing is done.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] StephanEwen commented on pull request #12481: [FLINK-18120] Don't expand documentation sections by default

2020-06-04 Thread GitBox


StephanEwen commented on pull request #12481:
URL: https://github.com/apache/flink/pull/12481#issuecomment-638791801


   +1
   
   This makes sense. The docs page currently shows too much contents 
immediately, which can be overwhelming.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] krasinski edited a comment on pull request #12462: [FLINK-18045] Fix Kerberos credentials checking

2020-06-04 Thread GitBox


krasinski edited a comment on pull request #12462:
URL: https://github.com/apache/flink/pull/12462#issuecomment-638494158


   thank you for the review @walterddr, I've addressed all the comments and 
added some more unit tests
   
   I've also introduced a small change in `hasHDFSDelegationToken` method - 
taking advantage of the UGI parameter instead of retrieving it one more time 
seems more natural
   
   edit: need to find another way to init kerberos config, tests worked on CI 
but failed on CI



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18122) Kubernetes test fails with "error: timed out waiting for the condition on jobs/flink-job-cluster"

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125855#comment-17125855
 ] 

Robert Metzger commented on FLINK-18122:


Ah, good catch. 
The test should have a better error handling to fail if the docker image build 
failed?

> Kubernetes test fails with "error: timed out waiting for the condition on 
> jobs/flink-job-cluster"
> -
>
> Key: FLINK-18122
> URL: https://issues.apache.org/jira/browse/FLINK-18122
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2697=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-06-04T09:25:40.7205843Z service/flink-job-cluster created
> 2020-06-04T09:25:40.9661515Z job.batch/flink-job-cluster created
> 2020-06-04T09:25:41.2189123Z deployment.apps/flink-task-manager created
> 2020-06-04T10:32:32.6402983Z error: timed out waiting for the condition on 
> jobs/flink-job-cluster
> 2020-06-04T10:32:33.8057757Z error: unable to upgrade connection: container 
> not found ("flink-task-manager")
> 2020-06-04T10:32:33.8111302Z sort: cannot read: 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*':
>  No such file or directory
> 2020-06-04T10:32:33.8124455Z FAIL WordCount: Output hash mismatch.  Got 
> d41d8cd98f00b204e9800998ecf8427e, expected e682ec6622b5e83f2eb614617d5ab2cf.
> 2020-06-04T10:32:33.8125379Z head hexdump of actual:
> 2020-06-04T10:32:33.8136133Z head: cannot open 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*'
>  for reading: No such file or directory
> 2020-06-04T10:32:33.8344715Z Debugging failed Kubernetes test:
> 2020-06-04T10:32:33.8345469Z Currently existing Kubernetes resources
> 2020-06-04T10:32:36.4977853Z I0604 10:32:36.497383   13191 request.go:621] 
> Throttling request took 1.198606989s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:46.6975735Z I0604 10:32:46.697234   13191 request.go:621] 
> Throttling request took 4.398107353s, request: 
> GET:https://10.1.0.4:8443/apis/authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:57.4978637Z I0604 10:32:57.497209   13191 request.go:621] 
> Throttling request took 1.198449167s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:07.4980104Z I0604 10:33:07.497320   13191 request.go:621] 
> Throttling request took 4.198274438s, request: 
> GET:https://10.1.0.4:8443/apis/apiextensions.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:18.4976060Z I0604 10:33:18.497258   13191 request.go:621] 
> Throttling request took 1.19871495s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:28.4979129Z I0604 10:33:28.497276   13191 request.go:621] 
> Throttling request took 4.198369672s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:30.9182069Z NAME READY   
> STATUS  RESTARTS   AGE
> 2020-06-04T10:33:30.9184099Z pod/flink-job-cluster-dtb67  0/1 
> ErrImageNeverPull   0  67m
> 2020-06-04T10:33:30.9184869Z pod/flink-task-manager-74ccc9bd9-psqwm   0/1 
> ErrImageNeverPull   0  67m
> 2020-06-04T10:33:30.9185226Z 
> 2020-06-04T10:33:30.9185926Z NAMETYPE
> CLUSTER-IP  EXTERNAL-IP   PORT(S) 
>   AGE
> 2020-06-04T10:33:30.9186832Z service/flink-job-cluster   NodePort
> 10.111.92.199   
> 6123:32501/TCP,6124:31360/TCP,6125:30025/TCP,8081:30081/TCP   67m
> 2020-06-04T10:33:30.9187545Z service/kubernetes  ClusterIP   
> 10.96.0.1   443/TCP 
>   68m
> 2020-06-04T10:33:30.9187976Z 
> 2020-06-04T10:33:30.9188472Z NAME READY   
> UP-TO-DATE   AVAILABLE   AGE
> 2020-06-04T10:33:30.9189179Z deployment.apps/flink-task-manager   0/1 1   
>  0   67m
> 2020-06-04T10:33:30.9189508Z 
> 2020-06-04T10:33:30.9189815Z NAME   
> DESIRED   CURRENT   READY   AGE
> 2020-06-04T10:33:30.9190418Z replicaset.apps/flink-task-manager-74ccc9bd9   1 
> 1 0   67m
> 2020-06-04T10:33:30.9190662Z 
> 2020-06-04T10:33:30.9190891Z NAME  COMPLETIONS   
> DURATION   AGE
> 2020-06-04T10:33:30.9191423Z job.batch/flink-job-cluster   0/1  

[jira] [Created] (FLINK-18124) Add documentation for new FLIP-27 source interface

2020-06-04 Thread Stephan Ewen (Jira)
Stephan Ewen created FLINK-18124:


 Summary: Add documentation for new FLIP-27 source interface
 Key: FLINK-18124
 URL: https://issues.apache.org/jira/browse/FLINK-18124
 Project: Flink
  Issue Type: Sub-task
  Components: Documentation
Reporter: Stephan Ewen
Assignee: Stephan Ewen
 Fix For: 1.11.0


The documentation should be under {{Application Development / DataStream API / 
Data Sources }}.

We need to add sections about
  - Data Source Concepts
  - Data Source API
  - Connector Base (Split Reader Library)
  - Event Time and Watermarks




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12468: [FLINK-18075] Wrap the SerializationSchema in KafkaSerializationSchema in Kafka connector

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12468:
URL: https://github.com/apache/flink/pull/12468#issuecomment-638243685


   
   ## CI report:
   
   * 4ae1da7c5aee0ff73e4461f0d9235dd60857f1d7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2651)
 
   * 7bf026a4e23a89df952f6b361ab8fe435c0a612d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12326: [FLINK-16577] [table-planner-blink] Fix numeric type mismatch error in  column interval relmetadata

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12326:
URL: https://github.com/apache/flink/pull/12326#issuecomment-633786013


   
   ## CI report:
   
   * 0b14ce7cf01c6664a3a98c35dd1c494a7ad4fc32 UNKNOWN
   * 23311a70dcb1522127aaadf906875d536b64c166 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2171)
 
   * 8c198bc1bb0b6be3321a01898e09d7492169b47e Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2725)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #11541: [FLINK-15416][network] add task manager netty client retry mechenism

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #11541:
URL: https://github.com/apache/flink/pull/11541#issuecomment-604812212


   
   ## CI report:
   
   * b5ef19285000d692d6b1b8b0aa7a37ce35c216f5 UNKNOWN
   * baf2c55f7bb0b9df7bdef94083c947693b75eae2 UNKNOWN
   * e5419f41a18eaae05814837d679f58480a2cb94c UNKNOWN
   * 34a70a74c479f177784b5ba4983b50072c4f7e5f Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2702)
 
   * f64492a673ed412c925b42152c4ce469d4ccbb0c Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2715)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18125) Skip CI execution on documentation pull requests

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18125?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125858#comment-17125858
 ] 

Robert Metzger commented on FLINK-18125:


[~chesnay] I'm not sure if we can skip the entire CI execution, because I don't 
know if the tests contain something for the docs. If that's the case, we could 
at least skip e2e execution.

> Skip CI execution on documentation pull requests
> 
>
> Key: FLINK-18125
> URL: https://issues.apache.org/jira/browse/FLINK-18125
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System / Azure Pipelines
>Reporter: Robert Metzger
>Priority: Major
>
> In order to save some resources, we can skip the CI execution on 
> documentation-only changes (whole changeset).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12478: [FLINK-17869][task][checkpointing] Fix race condition when caling ChannelStateWriter.abort

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12478:
URL: https://github.com/apache/flink/pull/12478#issuecomment-638685986


   
   ## CI report:
   
   * bb036ae536e7705d8dfd8b8a5654198bbf4fa5c2 UNKNOWN
   * 420b7b3f645625b0b43f056440509cdd53bbd4ef Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2694)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18103) Test Pandas DataFrame and Flink Table conversion

2020-06-04 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125746#comment-17125746
 ] 

Dian Fu commented on FLINK-18103:
-

Verified the following:


Convert from a PyFlink Table to a Pandas DataFrame
 * batch mode of old planner and blink planner
 * stream mode with bounded data set of old planner and blink planner (only 
supports append stream at present)
 * test different data types such as float, array, row, etc.


Convert from a Pandas DataFrame to a PyFlink Table
 * batch mode of blink planner (doesn't support batch mode of old planner)
 * stream mode of old planner and blink planner
 * test that it could infer the table schema from the given pandas DataFrame
 * test DataFrames with NaN data

> Test Pandas DataFrame and Flink Table conversion
> 
>
> Key: FLINK-18103
> URL: https://issues.apache.org/jira/browse/FLINK-18103
> Project: Flink
>  Issue Type: Test
>  Components: API / Python
>Reporter: sunjincheng
>Assignee: Dian Fu
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.11.0
>
>
> 1) Test convert a Pandas DataFrame to Flink Table
> 2) Test convert a Flink Table to a Pandas DataFrame



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-18103) Test Pandas DataFrame and Flink Table conversion

2020-06-04 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18103?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu closed FLINK-18103.
---
Resolution: Done

> Test Pandas DataFrame and Flink Table conversion
> 
>
> Key: FLINK-18103
> URL: https://issues.apache.org/jira/browse/FLINK-18103
> Project: Flink
>  Issue Type: Test
>  Components: API / Python
>Reporter: sunjincheng
>Assignee: Dian Fu
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.11.0
>
>
> 1) Test convert a Pandas DataFrame to Flink Table
> 2) Test convert a Flink Table to a Pandas DataFrame



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] kl0u closed pull request #12463: [FLINK-18087][yarn] Fix uploading user artifact for Yarn job cluster

2020-06-04 Thread GitBox


kl0u closed pull request #12463:
URL: https://github.com/apache/flink/pull/12463


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-18087) Uploading user artifact for Yarn job cluster could not work

2020-06-04 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-18087.
--
Resolution: Fixed

Fixed on master with d56454a3a0abe9ba34d1fec27864863bbd8a7c5f
and on release-1.11 with f394995281f13c888303025735adb56cc952

> Uploading user artifact for Yarn job cluster could not work
> ---
>
> Key: FLINK-18087
> URL: https://issues.apache.org/jira/browse/FLINK-18087
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Yang Wang
>Assignee: Yang Wang
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0, 1.12.0
>
>
> In FLINK-17632, we add the support remote user jar. However, uploading user 
> artifact for Yarn job cluster is broken exceptionally. In the following code, 
> we should only upload local files. Now it has the contrary behavior.
> {code:java}
> // only upload local files
> if (Utils.isRemotePath(entry.getValue().filePath)) {
>Path localPath = new Path(entry.getValue().filePath);
>Tuple2 remoteFileInfo =
>  fileUploader.uploadLocalFileToRemote(localPath, entry.getKey());
>jobGraph.setUserArtifactRemotePath(entry.getKey(), 
> remoteFileInfo.f0.toString());
> }
> {code}
>  
> Another problem is the related tests {{testPerJobModeWithDistributedCache}} 
> does not fail because we do not fetch the artifact from remote 
> filesystem(i.e. HDFS). We directly get it from local file. It also needs to 
> be enhanced.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wuchong closed pull request #12381: [FLINK-17941][sql-client] Switching catalog or database doesn't work …

2020-06-04 Thread GitBox


wuchong closed pull request #12381:
URL: https://github.com/apache/flink/pull/12381


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on a change in pull request #12375: [FLINK-17017][runtime] Implements bulk allocation for physical slots

2020-06-04 Thread GitBox


zhuzhurk commented on a change in pull request #12375:
URL: https://github.com/apache/flink/pull/12375#discussion_r435126559



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/SlotPoolImpl.java
##
@@ -447,12 +452,40 @@ public void releaseSlot(@Nonnull SlotRequestId 
slotRequestId, @Nullable Throwabl
 
componentMainThreadExecutor.assertRunningInMainThread();
 
+   checkState(batchSlotTimeoutCheckEnabled, "batch slot timeout 
check is disabled unexpectedly.");
+
final PendingRequest pendingRequest = 
PendingRequest.createBatchRequest(slotRequestId, resourceProfile);
 
return requestNewAllocatedSlotInternal(pendingRequest)
.thenApply(Function.identity());
}
 
+   @Override
+   public CompletableFuture 
requestNewAllocatedSlotWithoutTimeout(

Review comment:
   Agreed it would be better to explicitly disable the timeout check.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12456: [FLINK-17113][sql-cli] Refactor view support in SQL Client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12456:
URL: https://github.com/apache/flink/pull/12456#issuecomment-638032908


   
   ## CI report:
   
   * b5b8172d4ccddf197a084d38e6dcef13ee27da8e Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2614)
 
   * 12e3932ed8d74a59c0d4c28441b336284a2dffeb UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12480: [hotfix] Remove redundant backslash in the log of GPUDriver

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12480:
URL: https://github.com/apache/flink/pull/12480#issuecomment-638732843


   
   ## CI report:
   
   * 0704f5331a783e4d3dbd7c404e4b80c27e7b6a60 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2707)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-637399261


   
   ## CI report:
   
   * d553feca300e2668e3e5c16c1c524d73acde141c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2627)
 
   * 4d1f156fd758a44f4c394a478b6f37bc2de61c86 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12481: [FLINK-18120] Don't expand documentation sections by default

2020-06-04 Thread GitBox


flinkbot commented on pull request #12481:
URL: https://github.com/apache/flink/pull/12481#issuecomment-638746420


   
   ## CI report:
   
   * 57276170065599217eaf7a77f451d74a9311a870 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12455:
URL: https://github.com/apache/flink/pull/12455#issuecomment-638007072


   
   ## CI report:
   
   * 7e9143f4531e9a7477a1e2848020d7c69644a3be Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2666)
 
   * 278da87279e640156951deef07ee9e4e24cab361 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18075) Kafka connector does not call open method of (de)serialization schema

2020-06-04 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek updated FLINK-18075:
-
Issue Type: Bug  (was: Improvement)

> Kafka connector does not call open method of (de)serialization schema
> -
>
> Key: FLINK-18075
> URL: https://issues.apache.org/jira/browse/FLINK-18075
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, 
> SequenceFile)
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Seth Wiesman
>Assignee: Dawid Wysakowicz
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The Kafka consumer and producer do not  call the open methods of plain 
> (De)SerializationSchema interfaces. Only the Keyed and Kafka specific 
> interfaces. The updated SQL implementations such as 
> AvroRowDataSeriailzationSchema use these methods and so SQL queries using 
> avro and kafka will fail in a null pointer exception. 
> cc [~aljoscha]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18073) AvroRowDataSerializationSchema is not always serializable

2020-06-04 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18073?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek updated FLINK-18073:
-
Issue Type: Bug  (was: Improvement)

> AvroRowDataSerializationSchema is not always serializable
> -
>
> Key: FLINK-18073
> URL: https://issues.apache.org/jira/browse/FLINK-18073
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile), Table 
> SQL / Ecosystem
>Reporter: Seth Wiesman
>Assignee: Jark Wu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0, 1.12.0
>
>
> AvroRowDeserializationSchema member runtimeConverter may hold a reference to 
> non serializable avro schema fields. This value should be transient and 
> generated inside of open. 
> {code:java}
> Caused by: java.io.NotSerializableException: 
> org.apache.avro.Schema$RecordSchema
>   at 
> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1184) 
> ~[?:1.8.0_252]
>   at java.io.ObjectOutputStream.writeArray(ObjectOutputStream.java:1378) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1174) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432) 
> ~[?:1.8.0_252]
>   at 
> java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178) 
> ~[?:1.8.0_252]
>   at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348) 
> ~[?:1.8.0_252]
>   at 
> org.apache.flink.util.InstantiationUtil.serializeObject(InstantiationUtil.java:586)
>  ~[flink-dist_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.api.java.ClosureCleaner.clean(ClosureCleaner.java:133) 
> ~[flink-dist_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.api.java.ClosureCleaner.clean(ClosureCleaner.java:126) 
> ~[flink-dist_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.api.java.ClosureCleaner.clean(ClosureCleaner.java:71) 
> ~[flink-dist_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer.(FlinkKafkaProducer.java:632)
>  ~[flink-sql-connector-kafka_2.12-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer.(FlinkKafkaProducer.java:522)
>  ~[flink-sql-connector-kafka_2.12-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer.(FlinkKafkaProducer.java:483)
>  ~[flink-sql-connector-kafka_2.12-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
>   at 
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSink.createKafkaProducer(KafkaDynamicSink.java:60)
>  ~[flink-sql-connector-kafka_2.12-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
>   at 
> org.apache.flink.streaming.connectors.kafka.table.KafkaDynamicSinkBase.getSinkRuntimeProvider(KafkaDynamicSinkBase.java:84)
>  ~[flink-sql-connector-kafka_2.12-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
>   at 
> org.apache.flink.table.planner.plan.nodes.common.CommonPhysicalSink.createSinkTransformation(CommonPhysicalSink.scala:69)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:108)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlanInternal(StreamExecSink.scala:43)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan(ExecNode.scala:58)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.table.planner.plan.nodes.exec.ExecNode.translateToPlan$(ExecNode.scala:56)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> org.apache.flink.table.planner.plan.nodes.physical.stream.StreamExecSink.translateToPlan(StreamExecSink.scala:43)
>  ~[flink-table-blink_2.12-1.11-stream.jar:1.11-stream]
>   at 
> 

[jira] [Updated] (FLINK-17935) Logs could not show up when deploying Flink on Yarn via "--executor"

2020-06-04 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas updated FLINK-17935:
---
Priority: Blocker  (was: Critical)

> Logs could not show up when deploying Flink on Yarn via "--executor"
> 
>
> Key: FLINK-17935
> URL: https://issues.apache.org/jira/browse/FLINK-17935
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / YARN
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Yang Wang
>Assignee: Kostas Kloudas
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> {code:java}
> ./bin/flink run -d -p 5 -e yarn-per-job 
> examples/streaming/WindowJoin.jar{code}
> When we use the {{-e/--executor}} to specify the deploy target to Yarn 
> per-job, the logs could not show up. The root cause is we do not set the 
> logging files in {{ExecutorCLI}}. We only do it in the 
> {{FlinkYarnSessionCli}}.
> If we use {{-m yarn-cluster}}, everything works well.
>  
> Maybe we should move the {{setLogConfigFileInConfig}} to 
> {{YarnClusterDescriptor}} to avoid this problem. cc [~kkl0u]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17824) "Resuming Savepoint" e2e stalls indefinitely

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125795#comment-17125795
 ] 

Robert Metzger commented on FLINK-17824:


Another instance on master (which makes the Hadoop 3.1.3 theory less likely) 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2696=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5

> "Resuming Savepoint" e2e stalls indefinitely 
> -
>
> Key: FLINK-17824
> URL: https://issues.apache.org/jira/browse/FLINK-17824
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Major
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> CI; 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1887=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=94459a52-42b6-5bfc-5d74-690b5d3c6de8
> {code}
> 2020-05-19T21:05:52.9696236Z 
> ==
> 2020-05-19T21:05:52.9696860Z Running 'Resuming Savepoint (file, async, scale 
> down) end-to-end test'
> 2020-05-19T21:05:52.9697243Z 
> ==
> 2020-05-19T21:05:52.9713094Z TEST_DATA_DIR: 
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-52970362751
> 2020-05-19T21:05:53.1194478Z Flink dist directory: 
> /home/vsts/work/1/s/flink-dist/target/flink-1.12-SNAPSHOT-bin/flink-1.12-SNAPSHOT
> 2020-05-19T21:05:53.2180375Z Starting cluster.
> 2020-05-19T21:05:53.9986167Z Starting standalonesession daemon on host 
> fv-az558.
> 2020-05-19T21:05:55.5997224Z Starting taskexecutor daemon on host fv-az558.
> 2020-05-19T21:05:55.6223837Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.0552482Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.9446865Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:59.0098434Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:06:00.0569710Z Dispatcher REST endpoint is up.
> 2020-05-19T21:06:07.7099937Z Job (a92a74de8446a80403798bb4806b73f3) is 
> running.
> 2020-05-19T21:06:07.7855906Z Waiting for job to process up to 200 records, 
> current progress: 114 records ...
> 2020-05-19T21:06:55.5755111Z 
> 2020-05-19T21:06:55.5756550Z 
> 
> 2020-05-19T21:06:55.5757225Z  The program finished with the following 
> exception:
> 2020-05-19T21:06:55.5757566Z 
> 2020-05-19T21:06:55.5765453Z org.apache.flink.util.FlinkException: Could not 
> stop with a savepoint job "a92a74de8446a80403798bb4806b73f3".
> 2020-05-19T21:06:55.5766873Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:485)
> 2020-05-19T21:06:55.5767980Z  at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:854)
> 2020-05-19T21:06:55.5769014Z  at 
> org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:477)
> 2020-05-19T21:06:55.5770052Z  at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:921)
> 2020-05-19T21:06:55.5771107Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:982)
> 2020-05-19T21:06:55.5772223Z  at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
> 2020-05-19T21:06:55.5773325Z  at 
> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:982)
> 2020-05-19T21:06:55.5774871Z Caused by: 
> java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5777183Z  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
> 2020-05-19T21:06:55.5778884Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-19T21:06:55.5779920Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:483)
> 2020-05-19T21:06:55.5781175Z  ... 6 more
> 2020-05-19T21:06:55.5782391Z Caused by: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5783885Z  at 
> org.apache.flink.runtime.scheduler.SchedulerBase.lambda$stopWithSavepoint$9(SchedulerBase.java:890)
> 2020-05-19T21:06:55.5784992Z  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
> 2020-05-19T21:06:55.5786492Z  

[jira] [Updated] (FLINK-17824) "Resuming Savepoint" e2e stalls indefinitely

2020-06-04 Thread Robert Metzger (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Metzger updated FLINK-17824:
---
Priority: Critical  (was: Major)

> "Resuming Savepoint" e2e stalls indefinitely 
> -
>
> Key: FLINK-17824
> URL: https://issues.apache.org/jira/browse/FLINK-17824
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> CI; 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1887=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=94459a52-42b6-5bfc-5d74-690b5d3c6de8
> {code}
> 2020-05-19T21:05:52.9696236Z 
> ==
> 2020-05-19T21:05:52.9696860Z Running 'Resuming Savepoint (file, async, scale 
> down) end-to-end test'
> 2020-05-19T21:05:52.9697243Z 
> ==
> 2020-05-19T21:05:52.9713094Z TEST_DATA_DIR: 
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-52970362751
> 2020-05-19T21:05:53.1194478Z Flink dist directory: 
> /home/vsts/work/1/s/flink-dist/target/flink-1.12-SNAPSHOT-bin/flink-1.12-SNAPSHOT
> 2020-05-19T21:05:53.2180375Z Starting cluster.
> 2020-05-19T21:05:53.9986167Z Starting standalonesession daemon on host 
> fv-az558.
> 2020-05-19T21:05:55.5997224Z Starting taskexecutor daemon on host fv-az558.
> 2020-05-19T21:05:55.6223837Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.0552482Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.9446865Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:59.0098434Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:06:00.0569710Z Dispatcher REST endpoint is up.
> 2020-05-19T21:06:07.7099937Z Job (a92a74de8446a80403798bb4806b73f3) is 
> running.
> 2020-05-19T21:06:07.7855906Z Waiting for job to process up to 200 records, 
> current progress: 114 records ...
> 2020-05-19T21:06:55.5755111Z 
> 2020-05-19T21:06:55.5756550Z 
> 
> 2020-05-19T21:06:55.5757225Z  The program finished with the following 
> exception:
> 2020-05-19T21:06:55.5757566Z 
> 2020-05-19T21:06:55.5765453Z org.apache.flink.util.FlinkException: Could not 
> stop with a savepoint job "a92a74de8446a80403798bb4806b73f3".
> 2020-05-19T21:06:55.5766873Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:485)
> 2020-05-19T21:06:55.5767980Z  at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:854)
> 2020-05-19T21:06:55.5769014Z  at 
> org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:477)
> 2020-05-19T21:06:55.5770052Z  at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:921)
> 2020-05-19T21:06:55.5771107Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:982)
> 2020-05-19T21:06:55.5772223Z  at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
> 2020-05-19T21:06:55.5773325Z  at 
> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:982)
> 2020-05-19T21:06:55.5774871Z Caused by: 
> java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5777183Z  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
> 2020-05-19T21:06:55.5778884Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-19T21:06:55.5779920Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:483)
> 2020-05-19T21:06:55.5781175Z  ... 6 more
> 2020-05-19T21:06:55.5782391Z Caused by: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5783885Z  at 
> org.apache.flink.runtime.scheduler.SchedulerBase.lambda$stopWithSavepoint$9(SchedulerBase.java:890)
> 2020-05-19T21:06:55.5784992Z  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
> 2020-05-19T21:06:55.5786492Z  at 
> java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
> 2020-05-19T21:06:55.5787601Z  at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
> 

[jira] [Closed] (FLINK-17918) LIMIT queries are failed when adding sleeping time of async checkpoint

2020-06-04 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-17918.
---
Resolution: Fixed

[hotfix][table-planner-blink] Store last watermark in state to emit on recovery 
for EventTimeProcessOperator
 - master (1.12.0): 550d4e168bdf068b2c4eadd51878abaa98f0bb3c
 - 1.11.0: 7c72335a3dbf2a8de7469d2bef8074d1e33284c9

[FLINK-17918][table-blink] Fix AppendOnlyTopNFunction shouldn't mutate list 
value of MapState
 - master (1.12.0): 5ab403dafa3bc5ee8654173fb543969dfdd746f8
 - 1.11.0: a290c8a3af4083fb32febfc7f955a9bdce08e4b8


> LIMIT queries are failed when adding sleeping time of async checkpoint
> --
>
> Key: FLINK-17918
> URL: https://issues.apache.org/jira/browse/FLINK-17918
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Piotr Nowojski
>Assignee: Jark Wu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> When we change the timing of operations (sleep after emit first record and 
> sleep for async operation of checkpoint) with this 
> [commit|https://github.com/apache/flink/commit/c05a0d865989c9959047cebcf2cd68b3838cc699],
>  the test 
> {{org.apache.flink.table.planner.runtime.stream.sql.AggregateITCase#testDifferentTypesSumWithRetract}}
>  in flink-table-planner-blink is failed. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12410: [FLINK-13782][table-api] Implement type strategies for IF ELSE expression

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12410:
URL: https://github.com/apache/flink/pull/12410#issuecomment-635946149


   
   ## CI report:
   
   * ce34ce1e7b16cf8c6bf359ffadbf5d3e96613947 UNKNOWN
   * 5e2c5e2f0d9babcc26b6223449b3ab6f8ac97c3e Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2692)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12436:
URL: https://github.com/apache/flink/pull/12436#issuecomment-637399261


   
   ## CI report:
   
   * d553feca300e2668e3e5c16c1c524d73acde141c Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2627)
 
   * 4d1f156fd758a44f4c394a478b6f37bc2de61c86 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2711)
 
   * 63831f0ba3dc2a6ec31acc7fb1cfd1b51288945b Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2721)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12435: [FLINK-18059] [sql-client] Fix create/drop catalog statement can not be executed in sql client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12435:
URL: https://github.com/apache/flink/pull/12435#issuecomment-637372360


   
   ## CI report:
   
   * b5ef005c90fb95a63c97d06c488adf5e4abe2972 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2626)
 
   * 223701abe70c0815eb1f35af7383727385ce2e90 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2720)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12455:
URL: https://github.com/apache/flink/pull/12455#issuecomment-638007072


   
   ## CI report:
   
   * 278da87279e640156951deef07ee9e4e24cab361 UNKNOWN
   * 76cf3b92819b6a77bf9c50b6d7674f5df3f4c8f7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2712)
 
   * ca75aac93b4b9716e3d1df62f878b8bd7a7c9aab Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2722)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18123) KubernetesClusterDescriptorTest.testDeployApplicationCluster:135->checkUpdatedConfigAndResourceSetting:194 expected: but was:

2020-06-04 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18123:
--

 Summary: 
KubernetesClusterDescriptorTest.testDeployApplicationCluster:135->checkUpdatedConfigAndResourceSetting:194
 expected: but was:
 Key: FLINK-18123
 URL: https://issues.apache.org/jira/browse/FLINK-18123
 Project: Flink
  Issue Type: Bug
  Components: Deployment / Kubernetes
Affects Versions: 1.12.0
Reporter: Robert Metzger


CI: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2696=logs=fc5181b0-e452-5c8f-68de-1097947f6483=62110053-334f-5295-a0ab-80dd7e2babbf

{code}
2020-06-04T10:17:00.2971211Z [ERROR] Failures: 
2020-06-04T10:17:00.2973074Z [ERROR]   
KubernetesClusterDescriptorTest.testDeployApplicationCluster:135->checkUpdatedConfigAndResourceSetting:194
 expected: but was:
2020-06-04T10:17:00.2973889Z [ERROR] Errors: 
2020-06-04T10:17:00.2974940Z [ERROR]   
KubernetesClusterDescriptorTest.testDeployHighAvailabilitySessionCluster:86->deploySessionCluster:181
 » ClusterDeployment
2020-06-04T10:17:00.2976255Z [ERROR]   
KubernetesClusterDescriptorTest.testDeploySessionCluster:77->deploySessionCluster:181
 » ClusterDeployment
2020-06-04T10:17:00.2978026Z [ERROR]   
KubernetesClusterDescriptorTest.testKillCluster:112->deploySessionCluster:181 
» ClusterDeployment
2020-06-04T10:17:00.2980261Z [ERROR]   
KubernetesResourceManagerTest.testCreateTaskManagerPodFailedAndRetry:306 » 
Execution
2020-06-04T10:17:00.2981587Z [ERROR]   
KubernetesResourceManagerTest.testDuplicatedPodAdded:400 » Execution 
java.lang...
2020-06-04T10:17:00.2982675Z [ERROR]   
KubernetesResourceManagerTest.testGetWorkerNodesFromPreviousAttempts:277 » 
Execution
2020-06-04T10:17:00.2983856Z [ERROR]   
KubernetesResourceManagerTest.testPodAddedBeforeCreateTaskManagerPodFutureComplete:485
 » Execution
2020-06-04T10:17:00.2984993Z [ERROR]   
KubernetesResourceManagerTest.testPodTerminatedBeforeAdded:425 » Execution 
jav...
2020-06-04T10:17:00.2986083Z [ERROR]   
KubernetesResourceManagerTest.testPreviousAttemptPodAdded:368 » Execution 
java...
2020-06-04T10:17:00.2987522Z [ERROR]   
KubernetesResourceManagerTest.testStartAndRecoverVariousResourceSpec:328 » 
NullPointer
2020-06-04T10:17:00.2988812Z [ERROR]   
KubernetesResourceManagerTest.testStartAndStopWorker:188 » Execution 
java.lang...
2020-06-04T10:17:00.2989909Z [ERROR]   
KubernetesResourceManagerTest.testTaskManagerPodTerminated:236 » Execution 
jav...
2020-06-04T10:17:00.2990987Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2992098Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2993175Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2994271Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2995381Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2996474Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2997869Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.2999050Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.3000190Z [ERROR]   
Fabric8FlinkKubeClientTest>KubernetesTestBase.setup:101->onSetup:102 » 
NullPointer
2020-06-04T10:17:00.3001184Z [ERROR]   
FlinkConfMountDecoratorTest.testConfigMap:84 » NullPointer
2020-06-04T10:17:00.3002153Z [ERROR]   
FlinkConfMountDecoratorTest.testDecoratedFlinkContainer:214 » NullPointer
2020-06-04T10:17:00.3003179Z [ERROR]   
FlinkConfMountDecoratorTest.testDecoratedFlinkPodWithLog4j:133 » NullPointer
2020-06-04T10:17:00.3004271Z [ERROR]   
FlinkConfMountDecoratorTest.testDecoratedFlinkPodWithLog4jAndLogback:186 » 
NullPointer
2020-06-04T10:17:00.3005594Z [ERROR]   
FlinkConfMountDecoratorTest.testDecoratedFlinkPodWithLogback:159 » NullPointer
2020-06-04T10:17:00.3006692Z [ERROR]   
FlinkConfMountDecoratorTest.testDecoratedFlinkPodWithoutLog4jAndLogback:104 » 
NullPointer
2020-06-04T10:17:00.3007991Z [ERROR]   
FlinkConfMountDecoratorTest.testWhetherPodOrContainerIsDecorated:74 » 
NullPointer
2020-06-04T10:17:00.3009236Z [ERROR]   
JavaCmdJobManagerDecoratorTest.testContainerStartCommandTemplate1:200 » 
NullPointer
2020-06-04T10:17:00.3010379Z [ERROR]   
JavaCmdJobManagerDecoratorTest.testContainerStartCommandTemplate2:228 » 
NullPointer
2020-06-04T10:17:00.3011431Z [ERROR]   
JavaCmdJobManagerDecoratorTest.testStartCommandWithLog4j:114 » NullPointer
2020-06-04T10:17:00.3012454Z [ERROR]   
JavaCmdJobManagerDecoratorTest.testStartCommandWithLog4jAndLogback:143 » 
NullPointer

[GitHub] [flink] pnowojski commented on a change in pull request #12460: [FLINK-18063][checkpointing] Fix the race condition for aborting current checkpoint in CheckpointBarrierUnaligner

2020-06-04 Thread GitBox


pnowojski commented on a change in pull request #12460:
URL: https://github.com/apache/flink/pull/12460#discussion_r435188801



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/io/CheckpointBarrierUnaligner.java
##
@@ -154,46 +154,51 @@ public void processBarrier(CheckpointBarrier 
receivedBarrier, int channelIndex)
@Override
public void processCancellationBarrier(CancelCheckpointMarker 
cancelBarrier) throws Exception {
long cancelledId = cancelBarrier.getCheckpointId();
-   if (LOG.isDebugEnabled()) {
-   LOG.debug("{}: Checkpoint {} canceled, aborting 
alignment.", taskName, cancelledId);
-   }
-
-   if (currentConsumedCheckpointId >= cancelledId && 
!isCheckpointPending()) {
-   return;
-   }
+   // tag whether we should abort checkpoint from task thread view
+   boolean shouldAbort1 = false;
 
-   if (isCheckpointPending()) {
+   if (cancelledId > currentConsumedCheckpointId) {
+   currentConsumedCheckpointId = cancelledId;
+   shouldAbort1 = true;
+   } else if (cancelledId == currentConsumedCheckpointId && 
isCheckpointPending()) {
LOG.warn("{}: Received cancellation barrier for 
checkpoint {} before completing current checkpoint {}. " +
"Skipping current checkpoint.",
taskName,
cancelledId,
currentConsumedCheckpointId);
 
resetConsumedBarriers();
+   shouldAbort1 = true;
+   }
+
+   // tag whether we should abort checkpoint from 
threadSafeUnaligner view
+   boolean shouldAbort2 = 
threadSafeUnaligner.setCancelledCheckpointId(cancelledId);

Review comment:
   > If shouldAbort2 true and shouldAbort1 false, that means the 
notifyBarrierReceived triggered by netty thread happen earlier.
   
   But this `notifyBarrierReceived` from netty thread, would enqueue mailbox 
action. 
   
   If the mailbox action hasn't yet been executed, we could mark this 
checkpoint as aborted here, in this method (task thread), and prevent mail from 
executing (task thread).
   
   If the mailbox action has already executed (task thread), it could have left 
the `currentConsumedCheckpointId` field already up to date, so this method 
would already know whether to abort or not from `shouldAbort1`.
   
   Or is it about aborting the checkpoint after `notifyBarrierReceived` from 
netty thread, but before enqueued mailbox action was executed?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rkhachatryan commented on pull request #12457: [FLINK-18050][task][checkpointing] Fix double buffer recycling

2020-06-04 Thread GitBox


rkhachatryan commented on pull request #12457:
URL: https://github.com/apache/flink/pull/12457#issuecomment-638795603


   CI failure unrelated (`Resuming Savepoint` end-to-end test failed with 
timeout - FLINK-17824).
   
   Succeeded on my private branch: 
https://dev.azure.com/khachatryanroman/flink/_build/results?buildId=86=results.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18125) Skip CI execution on documentation pull requests

2020-06-04 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18125:
--

 Summary: Skip CI execution on documentation pull requests
 Key: FLINK-18125
 URL: https://issues.apache.org/jira/browse/FLINK-18125
 Project: Flink
  Issue Type: Improvement
  Components: Build System / Azure Pipelines
Reporter: Robert Metzger


In order to save some resources, we can skip the CI execution on 
documentation-only changes (whole changeset).





--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125806#comment-17125806
 ] 

Etienne Chauchot commented on FLINK-17961:
--

[~chesnay] ES source can definitely mask the overall complexity to the user. As 
an example in Apache Beam ([available 
here|https://github.com/apache/beam/blob/e1963c11f9a853564d62f83993dec08ed8a9321f/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java#L156])
 what we do is we use sliced scroll to split the input collection for parallel 
reading and apply it to the user ES query or to a default _select * from index_ 
when there is no provided query. Thus, the user API remains simple with 
_ESIO.read().from(index).withQuery(query)._ My worries here are more related to 
streaming and failover capabilities raised by Aljoscha. Even though ES is a 
main source (not an enrichment one IMO) it does not meet some Flink 
expectancies (cf comments above). So the question is reduced to: is it worth 
investing some time to make an ES source still? Regarding the thread on an ES 
table source, I'll read it and comment if I have anything useful to say.

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #11541: [FLINK-15416][network] add task manager netty client retry mechenism

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #11541:
URL: https://github.com/apache/flink/pull/11541#issuecomment-604812212


   
   ## CI report:
   
   * b5ef19285000d692d6b1b8b0aa7a37ce35c216f5 UNKNOWN
   * baf2c55f7bb0b9df7bdef94083c947693b75eae2 UNKNOWN
   * e5419f41a18eaae05814837d679f58480a2cb94c UNKNOWN
   * a5dc1df6976661f4dc56c5e739b54dd90efc660d Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/169527021) Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2679)
 
   * 34a70a74c479f177784b5ba4983b50072c4f7e5f Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2702)
 
   * f64492a673ed412c925b42152c4ce469d4ccbb0c Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2715)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12352: [FLINK-17717][sql-parser] Throws for DDL create temporary system func…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12352:
URL: https://github.com/apache/flink/pull/12352#issuecomment-634517986


   
   ## CI report:
   
   * 6900843764e0fe637babcb429194ef620be73b50 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2278)
 
   * 7dc817125d0684987814d45192eab87a41ebc5b5 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2716)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12444:
URL: https://github.com/apache/flink/pull/12444#issuecomment-637596755


   
   ## CI report:
   
   * 16ec8b694dc07564b09b60040d20b0da9c0e80c4 UNKNOWN
   * b2e066ffdb1333b0b14aacc5955e4d0841a57096 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2690)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12456: [FLINK-17113][sql-cli] Refactor view support in SQL Client

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12456:
URL: https://github.com/apache/flink/pull/12456#issuecomment-638032908


   
   ## CI report:
   
   * 12e3932ed8d74a59c0d4c28441b336284a2dffeb Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2713)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12455:
URL: https://github.com/apache/flink/pull/12455#issuecomment-638007072


   
   ## CI report:
   
   * 278da87279e640156951deef07ee9e4e24cab361 UNKNOWN
   * 76cf3b92819b6a77bf9c50b6d7674f5df3f4c8f7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2712)
 
   * ca75aac93b4b9716e3d1df62f878b8bd7a7c9aab UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12475: [FLINK-18076][table sql / client] Sql client uses wrong class loader when parsing queries

2020-06-04 Thread GitBox


flinkbot edited a comment on pull request #12475:
URL: https://github.com/apache/flink/pull/12475#issuecomment-638593958


   
   ## CI report:
   
   * 41240cd524255ffc7cf63d2bc2ee9ff38f3abcbd UNKNOWN
   * fda75fa6db31dc6cc66f5b4fa7f86db3d4e26356 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2681)
 
   * 49c19b93edb74ef702c97a4742f4936145913f35 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2717)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18121) Support creating Docker image from local Flink distribution

2020-06-04 Thread Yang Wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18121?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125815#comment-17125815
 ] 

Yang Wang commented on FLINK-18121:
---

If this ticket is done, then FLINK-17857 could also benefit from this and will 
not have to use host machine address to fetch the flink.tar.gz.

> Support creating Docker image from local Flink distribution
> ---
>
> Key: FLINK-18121
> URL: https://issues.apache.org/jira/browse/FLINK-18121
> Project: Flink
>  Issue Type: Improvement
>  Components: Dockerfiles
>Affects Versions: docker-1.11.0.0
>Reporter: Till Rohrmann
>Priority: Major
>
> Currently, 
> https://github.com/apache/flink-docker/blob/dev-master/Dockerfile-debian.template
>  only supports to create a Docker image from a Flink distribution which is 
> hosted on a web server. I think it would be helpful if we could also create a 
> Docker image from a Flink distribution which is stored on one's local file 
> system. That way, one would not have to upload the file or start a web server 
> for serving it.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] zhijiangW commented on pull request #12460: [FLINK-18063][checkpointing] Fix the race condition for aborting current checkpoint in CheckpointBarrierUnaligner

2020-06-04 Thread GitBox


zhijiangW commented on pull request #12460:
URL: https://github.com/apache/flink/pull/12460#issuecomment-638773131


   @flinkbot run azure



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Etienne Chauchot closed FLINK-17961.

Release Note: Closing this because it duplicates: 
https://issues.apache.org/jira/browse/FLINK-16713 and the above ticket refers a 
design doc.
  Resolution: Duplicate

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Reopened] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Etienne Chauchot reopened FLINK-17961:
--

Move comment to the proper location (not release notes)

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Etienne Chauchot closed FLINK-17961.

Resolution: Duplicate

Closing this because it duplicates: 
https://issues.apache.org/jira/browse/FLINK-16713 and the above ticket refers a 
design doc.

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17961) Create an Elasticsearch source

2020-06-04 Thread Etienne Chauchot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Etienne Chauchot updated FLINK-17961:
-
Release Note:   (was: Closing this because it duplicates: 
https://issues.apache.org/jira/browse/FLINK-16713 and the above ticket refers a 
design doc.)

> Create an Elasticsearch source
> --
>
> Key: FLINK-17961
> URL: https://issues.apache.org/jira/browse/FLINK-17961
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / ElasticSearch
>Reporter: Etienne Chauchot
>Priority: Minor
>
> There is only an Elasticsearch sink available. There are opensource github 
> repos such as [this 
> one|[https://github.com/mnubo/flink-elasticsearch-source-connector]]. Also 
> the apache bahir project does not provide an Elasticsearch source connector 
> for flink either. IMHO I think the project would benefit from having an 
> bundled source connector for ES alongside with the available sink connector.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-18020) SQLClientKafkaITCase.testKafka failed on Travis

2020-06-04 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek reassigned FLINK-18020:


Assignee: Aljoscha Krettek

> SQLClientKafkaITCase.testKafka failed on Travis
> ---
>
> Key: FLINK-18020
> URL: https://issues.apache.org/jira/browse/FLINK-18020
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Table SQL / Client, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dawid Wysakowicz
>Assignee: Aljoscha Krettek
>Priority: Blocker
> Fix For: 1.11.0
>
>
> {code}
> [ERROR] Tests run: 3, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 
> 151.386 s <<< FAILURE! - in 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase
> [ERROR] testKafka[1: kafka-version:0.11 
> kafka-sql-version:.*kafka-0.11.jar](org.apache.flink.tests.util.kafka.SQLClientKafkaITCase)
>   Time elapsed: 49.811 s  <<< FAILURE!
> java.lang.AssertionError: 
> Expected: ["2018-03-12 08:00:00.000,Alice,This was a warning.,2,Success 
> constant folding.", "2018-03-12 09:00:00.000,Bob,This was another 
> warning.,1,Success constant folding.", "2018-03-12 09:00:00.000,Steve,This 
> was another info.,2,Success constant folding.", "2018-03-12 
> 09:00:00.000,Alice,This was a info.,1,Success constant folding."] in any order
>  but: Not matched: "2018-03-12 08:00:00.000,Alice,This was a 
> warning.,6,Success constant folding."
>   at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
>   at org.junit.Assert.assertThat(Assert.java:956)
>   at org.junit.Assert.assertThat(Assert.java:923)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.checkCsvResultFile(SQLClientKafkaITCase.java:241)
>   at 
> org.apache.flink.tests.util.kafka.SQLClientKafkaITCase.testKafka(SQLClientKafkaITCase.java:172)
> {code}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2323=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18071) CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI

2020-06-04 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125842#comment-17125842
 ] 

Robert Metzger commented on FLINK-18071:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2697=logs=0da23115-68bb-5dcd-192c-bd4c8adebde1=4ed44b66-cdd6-5dcf-5f6a-88b07dda665d

> CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI
> 
>
> Key: FLINK-18071
> URL: https://issues.apache.org/jira/browse/FLINK-18071
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> CI: 
> https://dev.azure.com/georgeryan1322/Flink/_build/results?buildId=330=logs=6e58d712-c5cc-52fb-0895-6ff7bd56c46b=f30a8e80-b2cf-535c-9952-7f521a4ae374
> {code}
> [ERROR] Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 8.795 
> s <<< FAILURE! - in 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase
> [ERROR] 
> test(org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase)
>   Time elapsed: 4.647 s  <<< FAILURE!
> java.lang.AssertionError: List did not contain expected sequence of 200 
> elements, but was: [152, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 
> 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 
> 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 
> 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 
> 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 
> 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 
> 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 
> 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 
> 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 
> 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 
> 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 
> 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 
> 198, 199]
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.failList(CoordinatorEventsExactlyOnceITCase.java:160)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.checkListContainsSequence(CoordinatorEventsExactlyOnceITCase.java:148)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.test(CoordinatorEventsExactlyOnceITCase.java:143)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18122) Kubernetes test fails with "error: timed out waiting for the condition on jobs/flink-job-cluster"

2020-06-04 Thread Chesnay Schepler (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125850#comment-17125850
 ] 

Chesnay Schepler commented on FLINK-18122:
--

The docker image build failed, and it likely re-used the image from a previous 
run:
{code}
Step 4/14 : RUN set -ex;   wget -nv -O /usr/local/bin/gosu 
"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$(dpkg 
--print-architecture)";   wget -nv -O /usr/local/bin/gosu.asc "
 ---> Running in eba9fe332456
+ dpkg --print-architecture
+ wget -nv -O /usr/local/bin/gosu 
https://github.com/tianon/gosu/releases/download/1.11/gosu-amd64
2020-06-04 09:25:20 
URL:https://github-production-release-asset-2e65be.s3.amazonaws.com/19708981/82e9dd00-d091-11e8-8734-a1caffcee352?X-Amz-Algorithm=AWS4-HMAC-SHA256=AKIAIWNJYAX4CSVEH53A%2F20200604%2Fus-east-1%2Fs3%2Faws4_request=20200604T092519Z=300=dd2292d73c781cad49cafc4064206ce385dca050762004be10d4ae5e1f5eaa9e=host_id=0_id=19708981=attachment%3B%20filename%3Dgosu-amd64=application%2Foctet-stream
 [2294944/2294944] -> "/usr/local/bin/gosu" [1]
+ dpkg --print-architecture
+ wget -nv -O /usr/local/bin/gosu.asc 
https://github.com/tianon/gosu/releases/download/1.11/gosu-amd64.asc
GnuTLS: Error in the pull function.
Unable to establish SSL connection.
The command '/bin/sh -c set -ex;   wget -nv -O /usr/local/bin/gosu 
"https://github.com/tianon/gosu/releases/download/$GOSU_VERSION/gosu-$(dpkg 
--print-architecture)";   wget -nv -O /usr/local/bin/gosu.asc "
~/work/1/s

{code}

> Kubernetes test fails with "error: timed out waiting for the condition on 
> jobs/flink-job-cluster"
> -
>
> Key: FLINK-18122
> URL: https://issues.apache.org/jira/browse/FLINK-18122
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2697=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-06-04T09:25:40.7205843Z service/flink-job-cluster created
> 2020-06-04T09:25:40.9661515Z job.batch/flink-job-cluster created
> 2020-06-04T09:25:41.2189123Z deployment.apps/flink-task-manager created
> 2020-06-04T10:32:32.6402983Z error: timed out waiting for the condition on 
> jobs/flink-job-cluster
> 2020-06-04T10:32:33.8057757Z error: unable to upgrade connection: container 
> not found ("flink-task-manager")
> 2020-06-04T10:32:33.8111302Z sort: cannot read: 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*':
>  No such file or directory
> 2020-06-04T10:32:33.8124455Z FAIL WordCount: Output hash mismatch.  Got 
> d41d8cd98f00b204e9800998ecf8427e, expected e682ec6622b5e83f2eb614617d5ab2cf.
> 2020-06-04T10:32:33.8125379Z head hexdump of actual:
> 2020-06-04T10:32:33.8136133Z head: cannot open 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*'
>  for reading: No such file or directory
> 2020-06-04T10:32:33.8344715Z Debugging failed Kubernetes test:
> 2020-06-04T10:32:33.8345469Z Currently existing Kubernetes resources
> 2020-06-04T10:32:36.4977853Z I0604 10:32:36.497383   13191 request.go:621] 
> Throttling request took 1.198606989s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:46.6975735Z I0604 10:32:46.697234   13191 request.go:621] 
> Throttling request took 4.398107353s, request: 
> GET:https://10.1.0.4:8443/apis/authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:57.4978637Z I0604 10:32:57.497209   13191 request.go:621] 
> Throttling request took 1.198449167s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:07.4980104Z I0604 10:33:07.497320   13191 request.go:621] 
> Throttling request took 4.198274438s, request: 
> GET:https://10.1.0.4:8443/apis/apiextensions.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:18.4976060Z I0604 10:33:18.497258   13191 request.go:621] 
> Throttling request took 1.19871495s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:28.4979129Z I0604 10:33:28.497276   13191 request.go:621] 
> Throttling request took 4.198369672s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:30.9182069Z NAME READY   
> STATUS  RESTARTS   AGE
> 2020-06-04T10:33:30.9184099Z pod/flink-job-cluster-dtb67  0/1 
> ErrImageNeverPull   0  67m
> 2020-06-04T10:33:30.9184869Z pod/flink-task-manager-74ccc9bd9-psqwm   0/1 
> ErrImageNeverPull   0  67m

[jira] [Commented] (FLINK-18122) Kubernetes test fails with "error: timed out waiting for the condition on jobs/flink-job-cluster"

2020-06-04 Thread Chesnay Schepler (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18122?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125860#comment-17125860
 ] 

Chesnay Schepler commented on FLINK-18122:
--

Yes; it should be enough to check for errors in 
{{flink-end-to-end-tests/test-scripts/common_docker.sh:build_image}}

> Kubernetes test fails with "error: timed out waiting for the condition on 
> jobs/flink-job-cluster"
> -
>
> Key: FLINK-18122
> URL: https://issues.apache.org/jira/browse/FLINK-18122
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2697=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5
> {code}
> 2020-06-04T09:25:40.7205843Z service/flink-job-cluster created
> 2020-06-04T09:25:40.9661515Z job.batch/flink-job-cluster created
> 2020-06-04T09:25:41.2189123Z deployment.apps/flink-task-manager created
> 2020-06-04T10:32:32.6402983Z error: timed out waiting for the condition on 
> jobs/flink-job-cluster
> 2020-06-04T10:32:33.8057757Z error: unable to upgrade connection: container 
> not found ("flink-task-manager")
> 2020-06-04T10:32:33.8111302Z sort: cannot read: 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*':
>  No such file or directory
> 2020-06-04T10:32:33.8124455Z FAIL WordCount: Output hash mismatch.  Got 
> d41d8cd98f00b204e9800998ecf8427e, expected e682ec6622b5e83f2eb614617d5ab2cf.
> 2020-06-04T10:32:33.8125379Z head hexdump of actual:
> 2020-06-04T10:32:33.8136133Z head: cannot open 
> '/home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-56335570120/out/kubernetes_wc_out*'
>  for reading: No such file or directory
> 2020-06-04T10:32:33.8344715Z Debugging failed Kubernetes test:
> 2020-06-04T10:32:33.8345469Z Currently existing Kubernetes resources
> 2020-06-04T10:32:36.4977853Z I0604 10:32:36.497383   13191 request.go:621] 
> Throttling request took 1.198606989s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:46.6975735Z I0604 10:32:46.697234   13191 request.go:621] 
> Throttling request took 4.398107353s, request: 
> GET:https://10.1.0.4:8443/apis/authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:32:57.4978637Z I0604 10:32:57.497209   13191 request.go:621] 
> Throttling request took 1.198449167s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:07.4980104Z I0604 10:33:07.497320   13191 request.go:621] 
> Throttling request took 4.198274438s, request: 
> GET:https://10.1.0.4:8443/apis/apiextensions.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:18.4976060Z I0604 10:33:18.497258   13191 request.go:621] 
> Throttling request took 1.19871495s, request: 
> GET:https://10.1.0.4:8443/apis/apps/v1?timeout=32s
> 2020-06-04T10:33:28.4979129Z I0604 10:33:28.497276   13191 request.go:621] 
> Throttling request took 4.198369672s, request: 
> GET:https://10.1.0.4:8443/apis/rbac.authorization.k8s.io/v1?timeout=32s
> 2020-06-04T10:33:30.9182069Z NAME READY   
> STATUS  RESTARTS   AGE
> 2020-06-04T10:33:30.9184099Z pod/flink-job-cluster-dtb67  0/1 
> ErrImageNeverPull   0  67m
> 2020-06-04T10:33:30.9184869Z pod/flink-task-manager-74ccc9bd9-psqwm   0/1 
> ErrImageNeverPull   0  67m
> 2020-06-04T10:33:30.9185226Z 
> 2020-06-04T10:33:30.9185926Z NAMETYPE
> CLUSTER-IP  EXTERNAL-IP   PORT(S) 
>   AGE
> 2020-06-04T10:33:30.9186832Z service/flink-job-cluster   NodePort
> 10.111.92.199   
> 6123:32501/TCP,6124:31360/TCP,6125:30025/TCP,8081:30081/TCP   67m
> 2020-06-04T10:33:30.9187545Z service/kubernetes  ClusterIP   
> 10.96.0.1   443/TCP 
>   68m
> 2020-06-04T10:33:30.9187976Z 
> 2020-06-04T10:33:30.9188472Z NAME READY   
> UP-TO-DATE   AVAILABLE   AGE
> 2020-06-04T10:33:30.9189179Z deployment.apps/flink-task-manager   0/1 1   
>  0   67m
> 2020-06-04T10:33:30.9189508Z 
> 2020-06-04T10:33:30.9189815Z NAME   
> DESIRED   CURRENT   READY   AGE
> 2020-06-04T10:33:30.9190418Z replicaset.apps/flink-task-manager-74ccc9bd9   1 
> 1 0   67m
> 2020-06-04T10:33:30.9190662Z 
> 2020-06-04T10:33:30.9190891Z NAME  COMPLETIONS   
> DURATION   AGE
> 2020-06-04T10:33:30.9191423Z 

[GitHub] [flink] zhuzhurk commented on a change in pull request #12375: [FLINK-17017][runtime] Implements bulk allocation for physical slots

2020-06-04 Thread GitBox


zhuzhurk commented on a change in pull request #12375:
URL: https://github.com/apache/flink/pull/12375#discussion_r435194969



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/slotpool/BulkSlotProviderImpl.java
##
@@ -0,0 +1,271 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.jobmaster.slotpool;
+
+import org.apache.flink.annotation.VisibleForTesting;
+import org.apache.flink.api.common.time.Time;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.clusterframework.types.ResourceProfile;
+import org.apache.flink.runtime.clusterframework.types.SlotProfile;
+import org.apache.flink.runtime.concurrent.ComponentMainThreadExecutor;
+import org.apache.flink.runtime.concurrent.FutureUtils;
+import org.apache.flink.runtime.jobmaster.SlotInfo;
+import org.apache.flink.runtime.jobmaster.SlotRequestId;
+import org.apache.flink.util.clock.Clock;
+import org.apache.flink.util.clock.SystemClock;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Optional;
+import java.util.Set;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.TimeoutException;
+import java.util.stream.Collectors;
+import java.util.stream.Stream;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * Default implementation of {@link BulkSlotProvider}.
+ */
+class BulkSlotProviderImpl implements BulkSlotProvider {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(BulkSlotProviderImpl.class);
+
+   private ComponentMainThreadExecutor componentMainThreadExecutor;
+
+   private final SlotSelectionStrategy slotSelectionStrategy;
+
+   private final SlotPool slotPool;
+
+   private final Clock clock;
+
+   private final PhysicalSlotRequestBulkTracker slotRequestBulkTracker;
+
+   BulkSlotProviderImpl(final SlotSelectionStrategy slotSelectionStrategy, 
final SlotPool slotPool) {
+   this(slotSelectionStrategy, slotPool, 
SystemClock.getInstance());
+   }
+
+   @VisibleForTesting
+   BulkSlotProviderImpl(
+   final SlotSelectionStrategy slotSelectionStrategy,
+   final SlotPool slotPool,
+   final Clock clock) {
+
+   this.slotSelectionStrategy = 
checkNotNull(slotSelectionStrategy);
+   this.slotPool = checkNotNull(slotPool);
+   this.clock = checkNotNull(clock);
+
+   this.slotRequestBulkTracker = new 
PhysicalSlotRequestBulkTracker(clock);
+
+   this.componentMainThreadExecutor = new 
ComponentMainThreadExecutor.DummyComponentMainThreadExecutor(
+   "Scheduler is not initialized with proper main thread 
executor. " +
+   "Call to BulkSlotProvider.start(...) 
required.");
+   }
+
+   @Override
+   public void start(final ComponentMainThreadExecutor mainThreadExecutor) 
{
+   this.componentMainThreadExecutor = mainThreadExecutor;
+   }
+
+   @Override
+   public CompletableFuture> 
allocatePhysicalSlots(
+   final Collection 
physicalSlotRequests,
+   final Time timeout) {
+
+   componentMainThreadExecutor.assertRunningInMainThread();
+
+   LOG.debug("Received {} slot requests.", 
physicalSlotRequests.size());
+
+   final PhysicalSlotRequestBulk slotRequestBulk = new 
PhysicalSlotRequestBulk(physicalSlotRequests);
+
+   final List> 
resultFutures = new ArrayList<>(physicalSlotRequests.size());
+   for (PhysicalSlotRequest request : physicalSlotRequests) {
+   final CompletableFuture 
resultFuture =
+   allocatePhysicalSlot(request, 
timeout).thenApply(result -> {
+   slotRequestBulk.markRequestFulfilled(
+   result.getSlotRequestId(),
+   

[jira] [Commented] (FLINK-17726) Scheduler should take care of tasks directly canceled by TaskManager

2020-06-04 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17726?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125632#comment-17125632
 ] 

Nicholas Jiang commented on FLINK-17726:


I have already discuss with [~trohrmann] offline. And considering that I have 
completed other issues I claimed, I would like to make a design about this in a 
few these days, and discuss with you [~zhuzh] about the design. Thanks for 
[~zhuzh] and [~trohrmann] explanation for the solution.

> Scheduler should take care of tasks directly canceled by TaskManager
> 
>
> Key: FLINK-17726
> URL: https://issues.apache.org/jira/browse/FLINK-17726
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Runtime / Task
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Zhu Zhu
>Assignee: Nicholas Jiang
>Priority: Critical
> Fix For: 1.12.0, 1.11.1
>
>
> JobManager will not trigger failure handling when receiving CANCELED task 
> update. 
> This is because CANCELED tasks are usually caused by another FAILED task. 
> These CANCELED tasks will be restarted by the failover process triggered  
> FAILED task.
> However, if a task is directly CANCELED by TaskManager due to its own runtime 
> issue, the task will not be recovered by JM and thus the job would hang.
> This is a potential issue and we should avoid it.
> A possible solution is to let JobManager treat tasks transitioning to 
> CANCELED from all states except from CANCELING as failed tasks. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17918) LIMIT queries are failed when adding sleeping time of async checkpoint

2020-06-04 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17918?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125637#comment-17125637
 ] 

Jark Wu commented on FLINK-17918:
-

FYI, I have updated the title and description of this issue and forked out a 
new issue FLINK-18118 for the remaining problems.

> LIMIT queries are failed when adding sleeping time of async checkpoint
> --
>
> Key: FLINK-17918
> URL: https://issues.apache.org/jira/browse/FLINK-17918
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Piotr Nowojski
>Assignee: Jark Wu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> When we change the timing of operations (sleep after emit first record and 
> sleep for async operation of checkpoint) with this 
> [commit|https://github.com/apache/flink/commit/c05a0d865989c9959047cebcf2cd68b3838cc699],
>  the test 
> {{org.apache.flink.table.planner.runtime.stream.sql.AggregateITCase#testDifferentTypesSumWithRetract}}
>  in flink-table-planner-blink is failed. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] rkhachatryan opened a new pull request #12478: [FLINK-17869][task][checkpointing] Fix race condition when caling ChannelStateWriter.abort

2020-06-04 Thread GitBox


rkhachatryan opened a new pull request #12478:
URL: https://github.com/apache/flink/pull/12478


   __This is a draft PR. It depends on #12457 and should not be merged before__
   
   ## What is the purpose of the change
   
   Before FLINK-17218 there were two problems:
   1. Exceeding `ChannelStateWriterImpl.maxCheckpoints`
   1. Race condition with out of order barriers (see 
be8fbcf506b8dd38e5425cf772a55f033f0962b0 of this PR)
   
   FLINK-17218 (24ff415f1b76392f75dea7c3538558d24fcb7058) fixed the first one 
but introduced a new race condition when netty thread modifies a 
ChannelStateWriter map used by the task thread.
   
   This PR reverts FLINK-17218 and addresses the issues above.
   
   ## Brief change log
   1. Revert 24ff415f1b76392f75dea7c3538558d24fcb7058 
`"[FLINK-17218][checkpointing] Ensuring that ChannelStateWriter aborts previous 
checkpoints before a new checkpoint`
   1. Abort channel state write if checkpoint is subsumed - using future 
callback and task thread - fix race condition
   1. Increase `ChannelStateWriterImpl.DEFAULT_MAX_CHECKPOINTS` - fix the 
original problem of `FLINK-17218`
   1. Ignore out of order checkpoints in `SubtaskCheckpointCoordinator` - fix a 
more subtle race condition
   1. Unignore `UnalignedCheckpointITCase`
   
   ## Verifying this change
   
   Unignore `UnalignedCheckpointITCase`.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-17726) Scheduler should take care of tasks directly canceled by TaskManager

2020-06-04 Thread Zhu Zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17726?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125653#comment-17125653
 ] 

Zhu Zhu commented on FLINK-17726:
-

Thanks for the updates. [~nicholasjiang] 
Looking forward to your design.

> Scheduler should take care of tasks directly canceled by TaskManager
> 
>
> Key: FLINK-17726
> URL: https://issues.apache.org/jira/browse/FLINK-17726
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Runtime / Task
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Zhu Zhu
>Assignee: Nicholas Jiang
>Priority: Critical
> Fix For: 1.12.0, 1.11.1
>
>
> JobManager will not trigger failure handling when receiving CANCELED task 
> update. 
> This is because CANCELED tasks are usually caused by another FAILED task. 
> These CANCELED tasks will be restarted by the failover process triggered  
> FAILED task.
> However, if a task is directly CANCELED by TaskManager due to its own runtime 
> issue, the task will not be recovered by JM and thus the job would hang.
> This is a potential issue and we should avoid it.
> A possible solution is to let JobManager treat tasks transitioning to 
> CANCELED from all states except from CANCELING as failed tasks. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] azagrebin commented on a change in pull request #12256: [FLINK-17018][runtime] Allocates slots in bulks for pipelined region scheduling

2020-06-04 Thread GitBox


azagrebin commented on a change in pull request #12256:
URL: https://github.com/apache/flink/pull/12256#discussion_r435061970



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/OneSlotPerExecutionSlotAllocator.java
##
@@ -0,0 +1,277 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.scheduler;
+
+import org.apache.flink.api.common.time.Time;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.clusterframework.types.SlotProfile;
+import org.apache.flink.runtime.concurrent.FutureUtils;
+import org.apache.flink.runtime.jobmanager.scheduler.Locality;
+import org.apache.flink.runtime.jobmaster.LogicalSlot;
+import org.apache.flink.runtime.jobmaster.SlotOwner;
+import org.apache.flink.runtime.jobmaster.SlotRequestId;
+import org.apache.flink.runtime.jobmaster.slotpool.PhysicalSlot;
+import org.apache.flink.runtime.jobmaster.slotpool.PhysicalSlotRequest;
+import org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot;
+import org.apache.flink.runtime.jobmaster.slotpool.SlotProvider;
+import org.apache.flink.runtime.scheduler.strategy.ExecutionVertexID;
+import org.apache.flink.runtime.taskmanager.TaskManagerLocation;
+import org.apache.flink.util.FlinkException;
+
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.concurrent.CompletableFuture;
+import java.util.function.Function;
+import java.util.stream.Collectors;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+import static org.apache.flink.util.Preconditions.checkState;
+
+/**
+ * This slot allocator will request one physical slot for each single 
execution vertex.
+ * The slots will be requested in bulks so that the {@link SlotProvider} can 
check
+ * whether this bulk of slot requests can be fulfilled at the same time.
+ * It has several limitations:
+ *
+ * 1. Slot sharing will be ignored.
+ *
+ * 2. Co-location constraints are not allowed.
+ *
+ * 3. Intra-bulk input location preferences will be ignored.
+ */
+public class OneSlotPerExecutionSlotAllocator extends 
AbstractExecutionSlotAllocator {
+
+   private static final Logger LOG = 
LoggerFactory.getLogger(OneSlotPerExecutionSlotAllocator.class);
+
+   private final SlotProvider slotProvider;
+
+   private final SlotOwner slotOwner;
+
+   private final boolean slotWillBeOccupiedIndefinitely;
+
+   private final Time allocationTimeout;
+
+   public OneSlotPerExecutionSlotAllocator(
+   final SlotProvider slotProvider,

Review comment:
   Could we already make `OneSlotPerExecutionSlotAllocator` depending on 
`BulkSlotProvider` if `Scheduler` extends `BulkSlotProvider`?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18116) E2E performance test

2020-06-04 Thread Aljoscha Krettek (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125663#comment-17125663
 ] 

Aljoscha Krettek commented on FLINK-18116:
--

Please use a meaningful Issue title, such as "Manually test E2E performance on 
Flink 1.11", if we need this issue at all.



> E2E performance test
> 
>
> Key: FLINK-18116
> URL: https://issues.apache.org/jira/browse/FLINK-18116
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Core, API / DataStream, API / State Processor, 
> Build System, Client / Job Submission
>Affects Versions: 1.11.0
>Reporter: Aihua Li
>Assignee: Aihua Li
>Priority: Blocker
>  Labels: release-testing
> Fix For: 1.11.0
>
>
> it's mainly to verify the performance don't less than 1.10 version by 
> checking the metrics of end-to-end performance test,such as qps,latency .



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18076) Sql client uses wrong class loader when parsing queries

2020-06-04 Thread Leonard Xu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17125668#comment-17125668
 ] 

Leonard Xu commented on FLINK-18076:


[~dwysakowicz] sure, thanks.

> Sql client uses wrong class loader when parsing queries
> ---
>
> Key: FLINK-18076
> URL: https://issues.apache.org/jira/browse/FLINK-18076
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dawid Wysakowicz
>Assignee: Leonard Xu
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Sql-client when parsing queries does not use the user class loader from 
> {{ExecutionContext}}. This makes it impossible to query any sources if the 
> dependencies are added with {{-j}} flag.
> In order to reproduce it try querying e.g. KafkaDynamicSource with
> {code}
> CREATE TABLE MyUserTable (
>f0 BIGINT
> ) WITH (
>   'connector' = 'kafka',   
>   'topic' = 'topic_name', -- required: topic name from which the table is read
>  -- required: specify the Kafka server connection string
>   'properties.bootstrap.servers' = 'localhost:9092',
>   -- required for Kafka source, optional for Kafka sink, specify consumer 
> group
>   'properties.group.id' = 'testGroup',
>   -- optional: valid modes are "earliest-offset", "latest-offset", 
> "group-offsets", "specific-offsets" or "timestamp"
> 'scan.startup.mode' = 'earliest-offset',
>'format' = 'avro'
> );
> SELECT * FROM MyUserTable;
> {code}
> It give exception:
> {code}
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Unexpected exception. This is a bug. Please consider filing an issue.
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:213)
> Caused by: org.apache.flink.table.client.gateway.SqlExecutionException: 
> Invalidate SQL statement.
>   at 
> org.apache.flink.table.client.cli.SqlCommandParser.parseBySqlParser(SqlCommandParser.java:95)
>   at 
> org.apache.flink.table.client.cli.SqlCommandParser.parse(SqlCommandParser.java:79)
>   at 
> org.apache.flink.table.client.cli.CliClient.parseCommand(CliClient.java:256)
>   at org.apache.flink.table.client.cli.CliClient.open(CliClient.java:212)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:142)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:114)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201)
> Caused by: org.apache.flink.table.api.ValidationException: Unable to create a 
> source for reading table 'default_catalog.default_database.MyUserTable'.
> Table options are:
> 'connector'='kafka'
> 'format'='avro'
> 'properties.bootstrap.servers'='localhost:9092'
> 'properties.group.id'='testGroup'
> 'scan.startup.mode'='earliest-offset'
> 'topic'='topic_name'
>   at 
> org.apache.flink.table.factories.FactoryUtil.createTableSource(FactoryUtil.java:125)
>   at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.buildTableScan(CatalogSourceTable.scala:135)
>   at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.scala:78)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3492)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2415)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2102)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2051)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:661)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:642)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3345)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:568)
>   at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:164)
>   at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:151)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:773)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:745)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:238)
>   at 
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:66)
>   at 
> 

[jira] [Assigned] (FLINK-18118) Some SQL Jobs with two input operators are loosing data with unaligned checkpoints

2020-06-04 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18118?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski reassigned FLINK-18118:
--

Assignee: Arvid Heise

> Some SQL Jobs with two input operators are loosing data with unaligned 
> checkpoints
> --
>
> Key: FLINK-18118
> URL: https://issues.apache.org/jira/browse/FLINK-18118
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Table SQL / Runtime
>Affects Versions: 1.11.0
>Reporter: Jark Wu
>Assignee: Arvid Heise
>Priority: Blocker
> Fix For: 1.11.0
>
>
> After trying to enable unaligned checkpoints by default, a lot of Blink 
> streaming SQL/Table API tests containing joins or set operations are throwing 
> errors that are indicating we are loosing some data (full records, without 
> deserialisation errors). Example errors:
> {noformat}
> [ERROR] Failures: 
> [ERROR]   JoinITCase.testFullJoinWithEqualPk:775 expected: 3,3, null,4, null,5)> but was:
> [ERROR]   JoinITCase.testStreamJoinWithSameRecord:391 expected: 1,1,1,1, 2,2,2,2, 2,2,2,2, 3,3,3,3, 3,3,3,3, 4,4,4,4, 4,4,4,4, 5,5,5,5, 
> 5,5,5,5)> but was:
> [ERROR]   SemiAntiJoinStreamITCase.testAntiJoin:352 expected:<0> but was:<1>
> [ERROR]   SetOperatorsITCase.testIntersect:55 expected: 2,2,Hello, 3,2,Hello world)> but was:
> [ERROR]   JoinITCase.testJoinPushThroughJoin:1272 expected: 2,1,Hello, 2,1,Hello world)> but was:
> {noformat}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] kl0u commented on pull request #12455: [FLINK-17935] Move setting yarn.log-config-file to YarnClusterClientFactory

2020-06-04 Thread GitBox


kl0u commented on pull request #12455:
URL: https://github.com/apache/flink/pull/12455#issuecomment-638692082


   Yep, this seems to be a mistake.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




<    1   2   3   4   5   6   7   >