[flink-web] branch asf-site updated: Rebuild website

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new 4a39d70  Rebuild website
4a39d70 is described below

commit 4a39d70eeb6a0bfc4d07d5c5d969155d7d386bee
Author: Arvid Heise 
AuthorDate: Fri Oct 22 20:53:40 2021 +0200

Rebuild website
---
 content/blog/feed.xml   | 4 ++--
 content/news/2021/09/29/release-1.14.0.html | 4 ++--
 2 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/content/blog/feed.xml b/content/blog/feed.xml
index b699e73..18bf32d 100644
--- a/content/blog/feed.xml
+++ b/content/blog/feed.xml
@@ -213,7 +213,7 @@ seamlessly switching over from one source to the 
other.

The motivating use case for the Hybrid Source was to read streams from tiered storage setups as if there was one stream that spans all tiers. For example, new data may land in Kafa and is eventually migrated to S3 (typically in compressed columnar format, for cost efficiency and performance). -The Hybrid Source can read this as one contiguous logical stream, starting with the the historic data on S3 +The Hybrid Source can read this as one contiguous logical stream, starting with the historic data on S3 and transitioning over to the more recent data in Kafka.

@@ -268,7 +268,7 @@ reduce the amount of in-flight data stored in unaliged checkpoints under backpre
-

Buffer Deloating acts as a complementary feature, or even alternative, to unaligned checkpoints. +

Buffer Debloating acts as a complementary feature, or even alternative, to unaligned checkpoints. Checkout the documentation; to see how to activate this feature.

diff --git a/content/news/2021/09/29/release-1.14.0.html b/content/news/2021/09/29/release-1.14.0.html index 7e506dc..ba241c0 100644 --- a/content/news/2021/09/29/release-1.14.0.html +++ b/content/news/2021/09/29/release-1.14.0.html @@ -344,7 +344,7 @@ seamlessly switching over from one source to the other. The motivating use case for the Hybrid Source was to read streams from tiered storage setups as if there was one stream that spans all tiers. For example, new data may land in Kafa and is eventually migrated to S3 (typically in compressed columnar format, for cost efficiency and performance). -The Hybrid Source can read this as one contiguous logical stream, starting with the the historic data on S3 +The Hybrid Source can read this as one contiguous logical stream, starting with the historic data on S3 and transitioning over to the more recent data in Kafka. @@ -399,7 +399,7 @@ reduce the amount of in-flight data stored in unaliged checkpoints under backpre -Buffer Deloating acts as a complementary feature, or even alternative, to unaligned checkpoints. +Buffer Debloating acts as a complementary feature, or even alternative, to unaligned checkpoints. Checkout the https://nightlies.apache.org/flink/flink-docs-release-1.14/docs/deployment/memory/network_mem_tuning/#the-buffer-debloating-mechanism";>documentation to see how to activate this feature.

[flink-web] branch asf-site updated: fix some typos

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new c143c1e  fix some typos
c143c1e is described below

commit c143c1ef7c5038a1f84302b26be35bb2cad26cc5
Author: Jing Ge 
AuthorDate: Wed Oct 13 16:39:13 2021 +0200

fix some typos
---
 _posts/2021-09-29-release-1.14.0.md | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/_posts/2021-09-29-release-1.14.0.md 
b/_posts/2021-09-29-release-1.14.0.md
index d104be9..7e89577 100644
--- a/_posts/2021-09-29-release-1.14.0.md
+++ b/_posts/2021-09-29-release-1.14.0.md
@@ -125,7 +125,7 @@ seamlessly switching over from one source to the other.
 The motivating use case for the Hybrid Source was to read streams from tiered 
storage setups as if there was one
 stream that spans all tiers. For example, new data may land in Kafa and is 
eventually
 migrated to S3 (typically in compressed columnar format, for cost efficiency 
and performance).
-The Hybrid Source can read this as one contiguous logical stream, starting 
with the the historic data on S3
+The Hybrid Source can read this as one contiguous logical stream, starting 
with the historic data on S3
 and transitioning over to the more recent data in Kafka.
 
 
@@ -181,7 +181,7 @@ reduce the amount of in-flight data stored in unaliged 
checkpoints under backpre
   
 
 
-Buffer Deloating acts as a complementary feature, or even alternative, to 
unaligned checkpoints.
+Buffer Debloating acts as a complementary feature, or even alternative, to 
unaligned checkpoints.
 Checkout the 
[documentation](https://nightlies.apache.org/flink/flink-docs-release-1.14/docs/deployment/memory/network_mem_tuning/#the-buffer-debloating-mechanism)
 to see how to activate this feature.
 


[flink] 02/03: [FLINK-21407][doc][formats] Add parquet formats to DataSet connectors doc

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a commit to branch release-1.13
in repository https://gitbox.apache.org/repos/asf/flink.git

commit 499b55b50b1e79919cec672548211e244e8243e2
Author: Etienne Chauchot 
AuthorDate: Thu Oct 14 12:30:37 2021 +0200

[FLINK-21407][doc][formats] Add parquet formats to DataSet connectors doc
---
 .../docs/connectors/dataset/formats/parquet.md | 66 ++
 docs/content/docs/connectors/dataset/overview.md   |  1 +
 2 files changed, 67 insertions(+)

diff --git a/docs/content/docs/connectors/dataset/formats/parquet.md 
b/docs/content/docs/connectors/dataset/formats/parquet.md
new file mode 100644
index 000..30c254d
--- /dev/null
+++ b/docs/content/docs/connectors/dataset/formats/parquet.md
@@ -0,0 +1,66 @@
+---
+title:  "Parquet"
+weight: 4
+type: docs
+aliases:
+  - /dev/batch/connectors/formats/parquet.html
+---
+
+
+
+# Parquet format
+
+Flink has extensive built-in support for [Apache 
Parquet](http://parquet.apache.org/). This allows to easily read from Parquet 
files with Flink. 
+Be sure to include the Flink Parquet dependency to the pom.xml of your project.
+
+```xml
+
+  org.apache.flink
+  flink-parquet_{{< scala_version >}}
+  {{< version >}}
+
+```
+
+In order to read data from a Parquet file, you have to specify one of the 
implementation of `ParquetInputFormat`. There are several depending on your 
needs:
+- `ParquetPojoInputFormat` to read POJOs from parquet files
+- `ParquetRowInputFormat` to read Flink `Rows` (column oriented records) from 
parquet files
+- `ParquetMapInputFormat` to read Map records (Map of nested Flink type 
objects) from parquet files
+- `ParquetAvroInputFormat` to read Avro Generic Records from parquet files
+
+
+**Example for ParquetRowInputFormat**:
+
+```java
+MessageType parquetSchema = // use parquet libs to provide the parquet schema 
file and parse it or extract it from the parquet files
+ParquetRowInputFormat parquetInputFormat = new ParquetRowInputFormat(new 
Path(filePath),  parquetSchema);
+// project only needed fields if suited to reduce the amount of data. Use: 
parquetSchema#selectFields(projectedFieldNames);
+DataSet input = env.createInput(parquetInputFormat);
+```
+
+**Example for ParquetAvroInputFormat**:
+
+```java
+MessageType parquetSchema = // use parquet libs to provide the parquet schema 
file and parse it or extract it from the parquet files
+ParquetAvroInputFormat parquetInputFormat = new ParquetAvroInputFormat(new 
Path(filePath),  parquetSchema);
+// project only needed fields if suited to reduce the amount of data. Use: 
parquetSchema#selectFields(projectedFieldNames);
+DataSet input = env.createInput(parquetInputFormat);
+```
+
+
diff --git a/docs/content/docs/connectors/dataset/overview.md 
b/docs/content/docs/connectors/dataset/overview.md
index 7bb2344..87f80884 100644
--- a/docs/content/docs/connectors/dataset/overview.md
+++ b/docs/content/docs/connectors/dataset/overview.md
@@ -40,5 +40,6 @@ Flink supports the following formats:
 Microsoft Azure 
tale storage format
 MongoDb 
format
 
+These formats are gradually replaced with new Flink Source API starting with 
Flink 1.14.0. 
 
 {{< top >}}


[flink] 03/03: [FLINK-21407][doc][formats] Add formats to DataStream connectors doc

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a commit to branch release-1.13
in repository https://gitbox.apache.org/repos/asf/flink.git

commit e3627738314f935ef8ee68988247150328bc525f
Author: Etienne Chauchot 
AuthorDate: Tue Oct 19 16:09:37 2021 +0200

[FLINK-21407][doc][formats] Add formats to DataStream connectors doc
---
 .../docs/connectors/datastream/formats/_index.md   |  23 
 .../docs/connectors/datastream/formats/avro.md |  61 ++
 .../datastream/formats/azure_table_storage.md  | 130 +
 .../docs/connectors/datastream/formats/hadoop.md   |  38 ++
 .../docs/connectors/datastream/formats/mongodb.md  |  33 ++
 .../docs/connectors/datastream/formats/parquet.md  |  67 +++
 .../content/docs/connectors/datastream/overview.md |   7 +-
 7 files changed, 357 insertions(+), 2 deletions(-)

diff --git a/docs/content/docs/connectors/datastream/formats/_index.md 
b/docs/content/docs/connectors/datastream/formats/_index.md
new file mode 100644
index 000..282fc69
--- /dev/null
+++ b/docs/content/docs/connectors/datastream/formats/_index.md
@@ -0,0 +1,23 @@
+---
+title: Formats
+bookCollapseSection: true
+weight: 2
+---
+
diff --git a/docs/content/docs/connectors/datastream/formats/avro.md 
b/docs/content/docs/connectors/datastream/formats/avro.md
new file mode 100644
index 000..65c0bf9
--- /dev/null
+++ b/docs/content/docs/connectors/datastream/formats/avro.md
@@ -0,0 +1,61 @@
+---
+title:  "Avro"
+weight: 4
+type: docs
+aliases:
+- /dev/connectors/formats/avro.html
+- /apis/streaming/connectors/formats/avro.html
+---
+
+
+
+# Avro formats
+
+Flink has extensive built-in support for [Apache 
Avro](http://avro.apache.org/). This allows to easily read from Avro files with 
Flink.
+Also, the serialization framework of Flink is able to handle classes generated 
from Avro schemas. Be sure to include the Flink Avro dependency to the pom.xml 
of your project.
+
+```xml
+
+  org.apache.flink
+  flink-avro
+  {{< version >}}
+
+```
+
+In order to read data from an Avro file, you have to specify an 
`AvroInputFormat`.
+
+**Example**:
+
+```java
+AvroInputFormat users = new AvroInputFormat(in, User.class);
+DataStream usersDS = env.createInput(users);
+```
+
+Note that `User` is a POJO generated by Avro. Flink also allows to perform 
string-based key selection of these POJOs. For example:
+
+```java
+usersDS.keyBy("name")
+```
+
+
+Note that using the `GenericData.Record` type is possible with Flink, but not 
recommended. Since the record contains the full schema, its very data intensive 
and thus probably slow to use.
+
+Flink's POJO field selection also works with POJOs generated from Avro. 
However, the usage is only possible if the field types are written correctly to 
the generated class. If a field is of type `Object` you can not use the field 
as a join or grouping key.
+Specifying a field in Avro like this `{"name": "type_double_test", "type": 
"double"},` works fine, however specifying it as a UNION-type with only one 
field (`{"name": "type_double_test", "type": ["double"]},`) will generate a 
field of type `Object`. Note that specifying nullable types (`{"name": 
"type_double_test", "type": ["null", "double"]},`) is possible!
diff --git 
a/docs/content/docs/connectors/datastream/formats/azure_table_storage.md 
b/docs/content/docs/connectors/datastream/formats/azure_table_storage.md
new file mode 100644
index 000..f5c02c4
--- /dev/null
+++ b/docs/content/docs/connectors/datastream/formats/azure_table_storage.md
@@ -0,0 +1,130 @@
+---
+title:  "Microsoft Azure table"
+weight: 4
+type: docs
+aliases:
+- /dev/connectors/formats/azure_table_storage.html
+- /apis/streaming/connectors/formats/azure_table_storage.html
+---
+
+
+# Microsoft Azure Table Storage format
+
+_Note: This example works starting from Flink 0.6-incubating_
+
+This example is using the `HadoopInputFormat` wrapper to use an existing 
Hadoop input format implementation for accessing [Azure's Table 
Storage](https://azure.microsoft.com/en-us/documentation/articles/storage-introduction/).
+
+1. Download and compile the `azure-tables-hadoop` project. The input format 
developed by the project is not yet available in Maven Central, therefore, we 
have to build the project ourselves.
+   Execute the following commands:
+
+```bash
+git clone https://github.com/mooso/azure-tables-hadoop.git
+cd azure-tables-hadoop
+mvn clean install
+```
+
+2. Setup a new Flink project using the quickstarts:
+
+```bash
+curl https://flink.apache.org/q/quickstart.sh | bash
+```
+
+3. Add the following dependencies (in the `` section) to your 
`pom.xml` file:
+
+```xml
+
+   org.apache.flink
+   flink-hadoop-compatibility{{< scala_version >}}
+   {{< version >}}
+
+
+ com.microsoft.hadoop
+ microsoft-hadoop-azure
+ 0.0.4
+
+```
+
+`flink-hadoop-compatibility` is a Flink package that provides the Hadoop input 
format wrappers.
+`microsoft-hadoop-azure` is adding the pro

[flink] branch release-1.13 updated (ea53e4a -> e362773)

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a change to branch release-1.13
in repository https://gitbox.apache.org/repos/asf/flink.git.


from ea53e4a  [FLINK-24318][table-planner-blink] Casting a number to 
boolean has different results between 'select' fields and 'where' condition
 new 85a7f18  [FLINK-21407][doc][formats] Split DataSet connectors page 
into different formats and create a formats sub-folder like in table api doc.
 new 499b55b  [FLINK-21407][doc][formats] Add parquet formats to DataSet 
connectors doc
 new e362773  [FLINK-21407][doc][formats] Add formats to DataStream 
connectors doc

The 3 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 docs/content/docs/connectors/dataset/_index.md | 23 ++
 .../{table => dataset}/formats/_index.md   |  0
 .../docs/connectors/dataset/formats/avro.md| 60 
 .../formats/azure_table_storage.md}| 69 ++
 .../docs/connectors/dataset/formats/hadoop.md  | 39 +++
 .../docs/connectors/dataset/formats/mongodb.md | 31 +
 .../docs/connectors/dataset/formats/parquet.md | 66 ++
 docs/content/docs/connectors/dataset/overview.md   | 45 
 .../docs/connectors/datastream/formats/_index.md   | 23 ++
 .../docs/connectors/datastream/formats/avro.md | 61 
 .../formats/azure_table_storage.md}| 81 --
 .../docs/connectors/datastream/formats/hadoop.md   | 38 ++
 .../docs/connectors/datastream/formats/mongodb.md  | 33 +
 .../docs/connectors/datastream/formats/parquet.md  | 67 ++
 .../content/docs/connectors/datastream/overview.md |  7 +-
 15 files changed, 508 insertions(+), 135 deletions(-)
 create mode 100644 docs/content/docs/connectors/dataset/_index.md
 copy docs/content/docs/connectors/{table => dataset}/formats/_index.md (100%)
 create mode 100644 docs/content/docs/connectors/dataset/formats/avro.md
 copy docs/content/docs/connectors/{dataset.md => 
dataset/formats/azure_table_storage.md} (63%)
 create mode 100644 docs/content/docs/connectors/dataset/formats/hadoop.md
 create mode 100644 docs/content/docs/connectors/dataset/formats/mongodb.md
 create mode 100644 docs/content/docs/connectors/dataset/formats/parquet.md
 create mode 100644 docs/content/docs/connectors/dataset/overview.md
 create mode 100644 docs/content/docs/connectors/datastream/formats/_index.md
 create mode 100644 docs/content/docs/connectors/datastream/formats/avro.md
 rename docs/content/docs/connectors/{dataset.md => 
datastream/formats/azure_table_storage.md} (57%)
 create mode 100644 docs/content/docs/connectors/datastream/formats/hadoop.md
 create mode 100644 docs/content/docs/connectors/datastream/formats/mongodb.md
 create mode 100644 docs/content/docs/connectors/datastream/formats/parquet.md


[flink] 01/03: [FLINK-21407][doc][formats] Split DataSet connectors page into different formats and create a formats sub-folder like in table api doc.

2021-10-22 Thread arvid
This is an automated email from the ASF dual-hosted git repository.

arvid pushed a commit to branch release-1.13
in repository https://gitbox.apache.org/repos/asf/flink.git

commit 85a7f18fa5d299b18776addc82c56fe691cc1e66
Author: Etienne Chauchot 
AuthorDate: Thu Oct 14 11:28:12 2021 +0200

[FLINK-21407][doc][formats] Split DataSet connectors page into different 
formats and create a formats sub-folder like in table api doc.
---
 docs/content/docs/connectors/dataset/_index.md | 23 
 .../docs/connectors/dataset/formats/_index.md  | 23 
 .../docs/connectors/dataset/formats/avro.md| 60 +++
 .../formats/azure_table_storage.md}| 69 ++
 .../docs/connectors/dataset/formats/hadoop.md  | 39 
 .../docs/connectors/dataset/formats/mongodb.md | 31 ++
 docs/content/docs/connectors/dataset/overview.md   | 44 ++
 7 files changed, 225 insertions(+), 64 deletions(-)

diff --git a/docs/content/docs/connectors/dataset/_index.md 
b/docs/content/docs/connectors/dataset/_index.md
new file mode 100644
index 000..3b0d8f6
--- /dev/null
+++ b/docs/content/docs/connectors/dataset/_index.md
@@ -0,0 +1,23 @@
+---
+title: DataSet Connectors
+bookCollapseSection: true
+weight: 1
+---
+
diff --git a/docs/content/docs/connectors/dataset/formats/_index.md 
b/docs/content/docs/connectors/dataset/formats/_index.md
new file mode 100644
index 000..8faac16
--- /dev/null
+++ b/docs/content/docs/connectors/dataset/formats/_index.md
@@ -0,0 +1,23 @@
+---
+title: Formats
+bookCollapseSection: true
+weight: 2
+---
+
\ No newline at end of file
diff --git a/docs/content/docs/connectors/dataset/formats/avro.md 
b/docs/content/docs/connectors/dataset/formats/avro.md
new file mode 100644
index 000..b0c5572
--- /dev/null
+++ b/docs/content/docs/connectors/dataset/formats/avro.md
@@ -0,0 +1,60 @@
+---
+title:  "Avro"
+weight: 4
+type: docs
+aliases:
+  - /dev/batch/connectors/formats/avro.html
+---
+
+
+
+# Avro formats
+
+Flink has extensive built-in support for [Apache 
Avro](http://avro.apache.org/). This allows to easily read from Avro files with 
Flink.
+Also, the serialization framework of Flink is able to handle classes generated 
from Avro schemas. Be sure to include the Flink Avro dependency to the pom.xml 
of your project.
+
+```xml
+
+  org.apache.flink
+  flink-avro
+  {{< version >}}
+
+```
+
+In order to read data from an Avro file, you have to specify an 
`AvroInputFormat`.
+
+**Example**:
+
+```java
+AvroInputFormat users = new AvroInputFormat(in, User.class);
+DataSet usersDS = env.createInput(users);
+```
+
+Note that `User` is a POJO generated by Avro. Flink also allows to perform 
string-based key selection of these POJOs. For example:
+
+```java
+usersDS.groupBy("name")
+```
+
+
+Note that using the `GenericData.Record` type is possible with Flink, but not 
recommended. Since the record contains the full schema, its very data intensive 
and thus probably slow to use.
+
+Flink's POJO field selection also works with POJOs generated from Avro. 
However, the usage is only possible if the field types are written correctly to 
the generated class. If a field is of type `Object` you can not use the field 
as a join or grouping key.
+Specifying a field in Avro like this `{"name": "type_double_test", "type": 
"double"},` works fine, however specifying it as a UNION-type with only one 
field (`{"name": "type_double_test", "type": ["double"]},`) will generate a 
field of type `Object`. Note that specifying nullable types (`{"name": 
"type_double_test", "type": ["null", "double"]},`) is possible!
diff --git a/docs/content/docs/connectors/dataset.md 
b/docs/content/docs/connectors/dataset/formats/azure_table_storage.md
similarity index 63%
rename from docs/content/docs/connectors/dataset.md
rename to docs/content/docs/connectors/dataset/formats/azure_table_storage.md
index a4ddca3..406949b 100644
--- a/docs/content/docs/connectors/dataset.md
+++ b/docs/content/docs/connectors/dataset/formats/azure_table_storage.md
@@ -1,9 +1,9 @@
 ---
-title:  "DataSet Connectors"
-weight: 3
+title:  "Microsoft Azure table"
+weight: 4
 type: docs
 aliases:
-  - /dev/batch/connectors.html
+- /dev/batch/connectors/formats/azure_table_storage.html
 ---
 
 
-# DataSet Connectors
-
-## Reading from and writing to file systems
-
-The Apache Flink project supports multiple [file systems]({{< ref 
"docs/deployment/filesystems/overview" >}}) that can be used as backing stores
-for input and output connectors. 
-
-## Connecting to other systems using Input/OutputFormat wrappers for Hadoop
-
-Apache Flink allows users to access many different systems as data sources or 
sinks.
-The system is designed for very easy extensibility. Similar to Apache Hadoop, 
Flink has the concept
-of so called `InputFormat`s and `OutputFormat`s.
-
-One implementation of these `InputFormat`s is the `HadoopInputFormat`. This is 
a wrapper that allows
-users t

[flink] branch master updated (d14ca21 -> 2cb4773)

2021-10-22 Thread chesnay
This is an automated email from the ASF dual-hosted git repository.

chesnay pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git.


from d14ca21  [FLINK-24560][build][yarn] Copy example jars in 
pre-integration-test phase
 add 2cb4773  [FLINK-24561][build] Build archetype jars in compile phase

No new revisions were added by this update.

Summary of changes:
 flink-quickstart/flink-quickstart-java/pom.xml  | 17 +
 flink-quickstart/flink-quickstart-scala/pom.xml | 17 +
 2 files changed, 34 insertions(+)


[flink] branch master updated (f070f0d -> d14ca21)

2021-10-22 Thread chesnay
This is an automated email from the ASF dual-hosted git repository.

chesnay pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git.


from f070f0d  [FLINK-24331][tests] Forbid to finish the TestEventSource if 
it has not sent at least one value
 add d14ca21  [FLINK-24560][build][yarn] Copy example jars in 
pre-integration-test phase

No new revisions were added by this update.

Summary of changes:
 flink-yarn-tests/pom.xml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


[flink] 01/02: [refactor][tests] Expanded toString for TestEventSource

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a commit to branch release-1.14
in repository https://gitbox.apache.org/repos/asf/flink.git

commit ccd002f7f7d683442e860246aeee581e908a0a13
Author: Anton Kalashnikov 
AuthorDate: Thu Oct 21 15:34:03 2021 +0200

[refactor][tests] Expanded toString for TestEventSource
---
 .../operators/lifecycle/event/OperatorFinishedEvent.java  | 15 +++
 1 file changed, 15 insertions(+)

diff --git 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
index ce9bf0b..becc6fb 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
@@ -50,9 +50,19 @@ public class OperatorFinishedEvent extends TestEvent {
 return receiveInfo.forUpstream(upstreamID).getOrDefault(upstreamIndex, 
-1L);
 }
 
+@Override
+public String toString() {
+return String.format("%s; receive=%s; sent=%d", super.toString(), 
receiveInfo, lastSent);
+}
+
 /** LastVertexDataInfo. */
 public static class LastVertexDataInfo implements Serializable {
 public final Map bySubtask = new HashMap<>();
+
+@Override
+public String toString() {
+return String.valueOf(bySubtask);
+}
 }
 
 /** LastReceivedVertexDataInfo. */
@@ -67,5 +77,10 @@ public class OperatorFinishedEvent extends TestEvent {
 return byUpstreamOperatorID.getOrDefault(upstreamID, new 
LastVertexDataInfo())
 .bySubtask;
 }
+
+@Override
+public String toString() {
+return String.valueOf(byUpstreamOperatorID);
+}
 }
 }


[flink] branch release-1.14 updated (9ed6ad8 -> 471d682)

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a change to branch release-1.14
in repository https://gitbox.apache.org/repos/asf/flink.git.


from 9ed6ad8  [FLINK-24468][runtime] Replaced submit by execute for 
submitting buffer size calculation to mailbox
 new ccd002f  [refactor][tests] Expanded toString for TestEventSource
 new 471d682  [FLINK-24331][tests] Forbid to finish the TestEventSource if 
it has not sent at least one value

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 .../operators/lifecycle/event/OperatorFinishedEvent.java  | 15 +++
 .../operators/lifecycle/graph/TestEventSource.java|  3 ++-
 2 files changed, 17 insertions(+), 1 deletion(-)


[flink] 02/02: [FLINK-24331][tests] Forbid to finish the TestEventSource if it has not sent at least one value

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a commit to branch release-1.14
in repository https://gitbox.apache.org/repos/asf/flink.git

commit 471d682eedabb06fe795adf2dc583ce0e6081ec0
Author: Anton Kalashnikov 
AuthorDate: Thu Oct 21 16:59:18 2021 +0200

[FLINK-24331][tests] Forbid to finish the TestEventSource if it has not 
sent at least one value
---
 .../flink/runtime/operators/lifecycle/graph/TestEventSource.java   | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
index 0d32de9..1badb6a 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
@@ -73,7 +73,8 @@ class TestEventSource extends 
RichSourceFunction
 public void run(SourceContext ctx) {
 long lastSent = 0;
 while (isRunning) {
-TestCommand cmd = scheduledCommands.poll();
+// Don't finish the source if it has not sent at least one value.
+TestCommand cmd = lastSent == 0 ? null : scheduledCommands.poll();
 if (cmd == FINISH_SOURCES) {
 ack(cmd);
 isRunning = false;


[flink] 02/02: [FLINK-24331][tests] Forbid to finish the TestEventSource if it has not sent at least one value

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git

commit f070f0d9ae19275b1e32c2587209d7d8ae87b57e
Author: Anton Kalashnikov 
AuthorDate: Thu Oct 21 16:59:18 2021 +0200

[FLINK-24331][tests] Forbid to finish the TestEventSource if it has not 
sent at least one value
---
 .../flink/runtime/operators/lifecycle/graph/TestEventSource.java   | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
index 0d32de9..1badb6a 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/graph/TestEventSource.java
@@ -73,7 +73,8 @@ class TestEventSource extends 
RichSourceFunction
 public void run(SourceContext ctx) {
 long lastSent = 0;
 while (isRunning) {
-TestCommand cmd = scheduledCommands.poll();
+// Don't finish the source if it has not sent at least one value.
+TestCommand cmd = lastSent == 0 ? null : scheduledCommands.poll();
 if (cmd == FINISH_SOURCES) {
 ack(cmd);
 isRunning = false;


[flink] 01/02: [refactor][tests] Expanded toString for TestEventSource

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git

commit 63bd6a382807f4f09c07567d82c2e112d63fe28c
Author: Anton Kalashnikov 
AuthorDate: Thu Oct 21 15:34:03 2021 +0200

[refactor][tests] Expanded toString for TestEventSource
---
 .../operators/lifecycle/event/OperatorFinishedEvent.java  | 15 +++
 1 file changed, 15 insertions(+)

diff --git 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
index ce9bf0b..becc6fb 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/runtime/operators/lifecycle/event/OperatorFinishedEvent.java
@@ -50,9 +50,19 @@ public class OperatorFinishedEvent extends TestEvent {
 return receiveInfo.forUpstream(upstreamID).getOrDefault(upstreamIndex, 
-1L);
 }
 
+@Override
+public String toString() {
+return String.format("%s; receive=%s; sent=%d", super.toString(), 
receiveInfo, lastSent);
+}
+
 /** LastVertexDataInfo. */
 public static class LastVertexDataInfo implements Serializable {
 public final Map bySubtask = new HashMap<>();
+
+@Override
+public String toString() {
+return String.valueOf(bySubtask);
+}
 }
 
 /** LastReceivedVertexDataInfo. */
@@ -67,5 +77,10 @@ public class OperatorFinishedEvent extends TestEvent {
 return byUpstreamOperatorID.getOrDefault(upstreamID, new 
LastVertexDataInfo())
 .bySubtask;
 }
+
+@Override
+public String toString() {
+return String.valueOf(byUpstreamOperatorID);
+}
 }
 }


[flink] branch master updated (86c9694 -> f070f0d)

2021-10-22 Thread dwysakowicz
This is an automated email from the ASF dual-hosted git repository.

dwysakowicz pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git.


from 86c9694  [FLINK-24019][build][dist] Package Scala APIs separately
 new 63bd6a3  [refactor][tests] Expanded toString for TestEventSource
 new f070f0d  [FLINK-24331][tests] Forbid to finish the TestEventSource if 
it has not sent at least one value

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 .../operators/lifecycle/event/OperatorFinishedEvent.java  | 15 +++
 .../operators/lifecycle/graph/TestEventSource.java|  3 ++-
 2 files changed, 17 insertions(+), 1 deletion(-)


[flink] branch master updated (7b63482 -> 86c9694)

2021-10-22 Thread chesnay
This is an automated email from the ASF dual-hosted git repository.

chesnay pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git.


from 7b63482  [FLINK-24575][table-runtime] Migrate 
TestRowDataCsvInputFormat to DeserializationSchemaFactory
 add 86c9694  [FLINK-24019][build][dist] Package Scala APIs separately

No new revisions were added by this update.

Summary of changes:
 .../pom.xml| 115 ++---
 .../src/main/resources/META-INF/NOTICE |  16 +++
 .../main/resources/META-INF/licenses/LICENSE.scala |   0
 flink-dist/pom.xml |  32 --
 flink-dist/src/main/assemblies/bin.xml |  10 ++
 flink-dist/src/main/resources/META-INF/NOTICE  |   8 --
 pom.xml|   1 +
 7 files changed, 105 insertions(+), 77 deletions(-)
 copy {flink-metrics/flink-metrics-influxdb => flink-dist-scala}/pom.xml (53%)
 create mode 100644 flink-dist-scala/src/main/resources/META-INF/NOTICE
 rename {flink-dist => 
flink-dist-scala}/src/main/resources/META-INF/licenses/LICENSE.scala (100%)