[jira] [Commented] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Arvid Heise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205992#comment-17205992
 ] 

Arvid Heise commented on FLINK-19027:
-

Thanks, [~dian.fu] , I'm digging in.

> UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel
>  failed because of test timeout
> 
>
> Key: FLINK-19027
> URL: https://issues.apache.org/jira/browse/FLINK-19027
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.12.0, 1.11.2
>Reporter: Dian Fu
>Assignee: Arvid Heise
>Priority: Major
>  Labels: test-stability
> Fix For: 1.12.0, 1.11.3
>
>
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=5789=logs=119bbba7-f5e3-5e08-e72d-09f1529665de=ec103906-d047-5b8a-680e-05fc000dfca9]
> {code}
> 2020-08-22T21:13:05.5315459Z [ERROR] 
> shouldPerformUnalignedCheckpointOnParallelRemoteChannel(org.apache.flink.test.checkpointing.UnalignedCheckpointITCase)
>   Time elapsed: 300.075 s  <<< ERROR!
> 2020-08-22T21:13:05.5316451Z org.junit.runners.model.TestTimedOutException: 
> test timed out after 300 seconds
> 2020-08-22T21:13:05.5317432Z  at sun.misc.Unsafe.park(Native Method)
> 2020-08-22T21:13:05.5317799Z  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2020-08-22T21:13:05.5318247Z  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2020-08-22T21:13:05.5318885Z  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
> 2020-08-22T21:13:05.5327035Z  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2020-08-22T21:13:05.5328114Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2020-08-22T21:13:05.5328869Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1719)
> 2020-08-22T21:13:05.5329482Z  at 
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:74)
> 2020-08-22T21:13:05.5330138Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1699)
> 2020-08-22T21:13:05.5330771Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1681)
> 2020-08-22T21:13:05.5331351Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.execute(UnalignedCheckpointITCase.java:158)
> 2020-08-22T21:13:05.5332015Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel(UnalignedCheckpointITCase.java:140)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Arvid Heise (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arvid Heise reassigned FLINK-19027:
---

Assignee: Arvid Heise

> UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel
>  failed because of test timeout
> 
>
> Key: FLINK-19027
> URL: https://issues.apache.org/jira/browse/FLINK-19027
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.12.0, 1.11.2
>Reporter: Dian Fu
>Assignee: Arvid Heise
>Priority: Major
>  Labels: test-stability
> Fix For: 1.12.0, 1.11.3
>
>
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=5789=logs=119bbba7-f5e3-5e08-e72d-09f1529665de=ec103906-d047-5b8a-680e-05fc000dfca9]
> {code}
> 2020-08-22T21:13:05.5315459Z [ERROR] 
> shouldPerformUnalignedCheckpointOnParallelRemoteChannel(org.apache.flink.test.checkpointing.UnalignedCheckpointITCase)
>   Time elapsed: 300.075 s  <<< ERROR!
> 2020-08-22T21:13:05.5316451Z org.junit.runners.model.TestTimedOutException: 
> test timed out after 300 seconds
> 2020-08-22T21:13:05.5317432Z  at sun.misc.Unsafe.park(Native Method)
> 2020-08-22T21:13:05.5317799Z  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2020-08-22T21:13:05.5318247Z  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2020-08-22T21:13:05.5318885Z  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
> 2020-08-22T21:13:05.5327035Z  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2020-08-22T21:13:05.5328114Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2020-08-22T21:13:05.5328869Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1719)
> 2020-08-22T21:13:05.5329482Z  at 
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:74)
> 2020-08-22T21:13:05.5330138Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1699)
> 2020-08-22T21:13:05.5330771Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1681)
> 2020-08-22T21:13:05.5331351Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.execute(UnalignedCheckpointITCase.java:158)
> 2020-08-22T21:13:05.5332015Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel(UnalignedCheckpointITCase.java:140)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] gm7y8 commented on a change in pull request #13458: FLINK-18851 [runtime] Add checkpoint type to checkpoint history entries in Web UI

2020-10-01 Thread GitBox


gm7y8 commented on a change in pull request #13458:
URL: https://github.com/apache/flink/pull/13458#discussion_r498619466



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/rest/messages/checkpoints/CheckpointStatistics.java
##
@@ -229,8 +242,8 @@ public static CheckpointStatistics 
generateCheckpointStatistics(AbstractCheckpoi
checkpointStatisticsPerTask.put(
taskStateStat.getJobVertexId(),
new TaskCheckpointStatistics(
-   
checkpointStats.getCheckpointId(),
-   
checkpointStats.getStatus(),
+   
checkpointStats.getCheckpointId(),
+   checkpointStats.getStatus(),

Review comment:
   Thanks for pointing out. let me fix this.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13375: [FLINK-18815][filesystems][tests] Join SafetyNetCloseableRegistry's reaper thread before assertion

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13375:
URL: https://github.com/apache/flink/pull/13375#issuecomment-691410175


   
   ## CI report:
   
   * ce25c6bbc562da022b86aaa99d24f40c1d159b41 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7160)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] gm7y8 commented on a change in pull request #13458: FLINK-18851 [runtime] Add checkpoint type to checkpoint history entries in Web UI

2020-10-01 Thread GitBox


gm7y8 commented on a change in pull request #13458:
URL: https://github.com/apache/flink/pull/13458#discussion_r498611476



##
File path: flink-runtime-web/src/test/resources/rest_api_v1.snapshot
##
@@ -557,7 +557,7 @@
   },
   "status" : {
 "type" : "string",
-"enum" : [ "CREATED", "RUNNING", "FAILING", "FAILED", 
"CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", "RECONCILING" ]
+"enum" : [ "INITIALIZING", "CREATED", "RUNNING", "FAILING", 
"FAILED", "CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", 
"RECONCILING" ]

Review comment:
   this was auto-generated code.. not related to this change.. will open 
jira issue and commit this change separately.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-playgrounds] shuiqiangchen commented on a change in pull request #16: [FLINK-19145][walkthroughs] Add PyFlink-walkthrough to Flink playground.

2020-10-01 Thread GitBox


shuiqiangchen commented on a change in pull request #16:
URL: https://github.com/apache/flink-playgrounds/pull/16#discussion_r498607377



##
File path: pyflink-walkthrough/README.md
##
@@ -0,0 +1,140 @@
+# pyflink-walkthrough
+
+## Background
+
+In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
+
+* Reading data from a Kafka source;
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Performing a simple aggregation over the source data;
+* Writing the results to Elasticsearch and visualizing them in Kibana.
+
+The environment is based on Docker Compose, so the only requirement is that 
you have [Docker](https://docs.docker.com/get-docker/) 
+installed in your machine.
+
+### Kafka
+You will be using Kafka to store sample input data about payment transactions. 
A simple data generator 
[generate_source_data.py](generator/generate_source_data.py) is provided to
+continuously write new records to the `payment_msg` Kafka topic. Each record 
is structured as follows:
+ 
+`{"createTime": "2020-08-12 06:29:02", "orderId": 1597213797, "payAmount": 
28306.44976403719, "payPlatform": 0, "provinceId": 4}`
+
+* `createTime`: The creation time of the transaction. 
+* `orderId`: The id of the current transaction.
+* `payAmount`: The amount being paid with this transaction.
+* `payPlatform`: The platform used to create this payment: pc or mobile.
+* `provinceId`: The id of the province for the user. 
+
+You can use the following command to read data from the Kafka topic and check 
whether it's generated correctly:
+```shell script
+$ docker-compose exec kafka kafka-console-consumer.sh --bootstrap-server 
kafka:9092 --topic payment_msg
+{"createTime":"2020-07-27 
09:25:32.77","orderId":1595841867217,"payAmount":7732.44,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.231","orderId":1595841867218,"payAmount":75774.05,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.72","orderId":1595841867219,"payAmount":65908.55,"payPlatform":0,"provinceId":0}
+{"createTime":"2020-07-27 
09:25:34.216","orderId":1595841867220,"payAmount":15341.11,"payPlatform":0,"provinceId":1}
+{"createTime":"2020-07-27 
09:25:34.698","orderId":1595841867221,"payAmount":37504.42,"payPlatform":0,"provinceId":0}
+```
+You can also create a new topic by executing the following command:
+```shell script
+$ docker-compose exec kafka kafka-topics.sh --bootstrap-server kafka:9092 
--create --topic  --partitions 8 --replication-factor 1
+```
+
+### PyFlink
+
+The transaction data will be processed with PyFlink using the Python script 
[payment_msg_processing.py](payment_msg_proccessing.py).
+This script will first map the `provinceId` in the input records to its 
corresponding province name using a Python UDF, 
+and then compute the sum of the transaction amounts for each province.
+
+### ElasticSearch
+
+ElasticSearch is used to store the results and to provide an efficient query 
service.
+
+### Kibana
+
+Kibana is an open source data visualization dashboard for ElasticSearch. You 
will use it to visualize 
+the results of your PyFlink pipeline.
+
+## Setup
+
+As mentioned, the environment for this walkthrough is based on Docker Compose; 
It uses a custom image
+to spin up Flink (JobManager+TaskManager), Kafka+Zookeeper, the data 
generator, and Elasticsearch+Kibana containers.
+
+You can find the [docker-compose.yaml](docker-compose.yml) file of the 
pyflink-walkthrough in the `pyflink-walkthrough` root directory.
+
+### Building the Docker image
+
+First, build the Docker image by running:
+
+```bash
+docker-compose build
+```
+
+### Starting the Playground
+
+Once the Docker image build is complete, run the following command to start 
the playground:
+
+```bash
+docker-compose up -d
+```
+
+One way of checking if the playground was successfully started is to access 
some of the services that are exposed:
+
+1. visiting Flink Web UI [http://localhost:8081](http://localhost:8081).
+2. visiting Elasticsearch [http://localhost:9200](http://localhost:9200).
+3. visiting Kibana [http://localhost:5601](http://localhost:5601).
+
+**Note:** you may need to wait around 1 minute before all the services come up.
+
+### Stopping the Playground
+
+To stop the playground, run the following command:
+
+```bash
+docker-compose down
+```
+
+
+## Running the PyFlink job
+
+1. Submit the PyFlink job.
+```shell script
+$ docker-compose exec jobmanager ./bin/flink run -py 
/opt/pyflink-walkthrough/payment_msg_proccessing.py -d
+```
+Navigate to the [Flink Web UI](http://localhost:8081) after the job is 
submitted successfully. There should be a job in the running job list.
+Click the job to get more details. You should see that the `StreamGraph` of 
the `payment_msg_proccessing` consists of two nodes, each with a parallelism of 
1. 

[GitHub] [flink-playgrounds] shuiqiangchen commented on a change in pull request #16: [FLINK-19145][walkthroughs] Add PyFlink-walkthrough to Flink playground.

2020-10-01 Thread GitBox


shuiqiangchen commented on a change in pull request #16:
URL: https://github.com/apache/flink-playgrounds/pull/16#discussion_r498607377



##
File path: pyflink-walkthrough/README.md
##
@@ -0,0 +1,140 @@
+# pyflink-walkthrough
+
+## Background
+
+In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
+
+* Reading data from a Kafka source;
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Performing a simple aggregation over the source data;
+* Writing the results to Elasticsearch and visualizing them in Kibana.
+
+The environment is based on Docker Compose, so the only requirement is that 
you have [Docker](https://docs.docker.com/get-docker/) 
+installed in your machine.
+
+### Kafka
+You will be using Kafka to store sample input data about payment transactions. 
A simple data generator 
[generate_source_data.py](generator/generate_source_data.py) is provided to
+continuously write new records to the `payment_msg` Kafka topic. Each record 
is structured as follows:
+ 
+`{"createTime": "2020-08-12 06:29:02", "orderId": 1597213797, "payAmount": 
28306.44976403719, "payPlatform": 0, "provinceId": 4}`
+
+* `createTime`: The creation time of the transaction. 
+* `orderId`: The id of the current transaction.
+* `payAmount`: The amount being paid with this transaction.
+* `payPlatform`: The platform used to create this payment: pc or mobile.
+* `provinceId`: The id of the province for the user. 
+
+You can use the following command to read data from the Kafka topic and check 
whether it's generated correctly:
+```shell script
+$ docker-compose exec kafka kafka-console-consumer.sh --bootstrap-server 
kafka:9092 --topic payment_msg
+{"createTime":"2020-07-27 
09:25:32.77","orderId":1595841867217,"payAmount":7732.44,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.231","orderId":1595841867218,"payAmount":75774.05,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.72","orderId":1595841867219,"payAmount":65908.55,"payPlatform":0,"provinceId":0}
+{"createTime":"2020-07-27 
09:25:34.216","orderId":1595841867220,"payAmount":15341.11,"payPlatform":0,"provinceId":1}
+{"createTime":"2020-07-27 
09:25:34.698","orderId":1595841867221,"payAmount":37504.42,"payPlatform":0,"provinceId":0}
+```
+You can also create a new topic by executing the following command:
+```shell script
+$ docker-compose exec kafka kafka-topics.sh --bootstrap-server kafka:9092 
--create --topic  --partitions 8 --replication-factor 1
+```
+
+### PyFlink
+
+The transaction data will be processed with PyFlink using the Python script 
[payment_msg_processing.py](payment_msg_proccessing.py).
+This script will first map the `provinceId` in the input records to its 
corresponding province name using a Python UDF, 
+and then compute the sum of the transaction amounts for each province.
+
+### ElasticSearch
+
+ElasticSearch is used to store the results and to provide an efficient query 
service.
+
+### Kibana
+
+Kibana is an open source data visualization dashboard for ElasticSearch. You 
will use it to visualize 
+the results of your PyFlink pipeline.
+
+## Setup
+
+As mentioned, the environment for this walkthrough is based on Docker Compose; 
It uses a custom image
+to spin up Flink (JobManager+TaskManager), Kafka+Zookeeper, the data 
generator, and Elasticsearch+Kibana containers.
+
+You can find the [docker-compose.yaml](docker-compose.yml) file of the 
pyflink-walkthrough in the `pyflink-walkthrough` root directory.
+
+### Building the Docker image
+
+First, build the Docker image by running:
+
+```bash
+docker-compose build
+```
+
+### Starting the Playground
+
+Once the Docker image build is complete, run the following command to start 
the playground:
+
+```bash
+docker-compose up -d
+```
+
+One way of checking if the playground was successfully started is to access 
some of the services that are exposed:
+
+1. visiting Flink Web UI [http://localhost:8081](http://localhost:8081).
+2. visiting Elasticsearch [http://localhost:9200](http://localhost:9200).
+3. visiting Kibana [http://localhost:5601](http://localhost:5601).
+
+**Note:** you may need to wait around 1 minute before all the services come up.
+
+### Stopping the Playground
+
+To stop the playground, run the following command:
+
+```bash
+docker-compose down
+```
+
+
+## Running the PyFlink job
+
+1. Submit the PyFlink job.
+```shell script
+$ docker-compose exec jobmanager ./bin/flink run -py 
/opt/pyflink-walkthrough/payment_msg_proccessing.py -d
+```
+Navigate to the [Flink Web UI](http://localhost:8081) after the job is 
submitted successfully. There should be a job in the running job list.
+Click the job to get more details. You should see that the `StreamGraph` of 
the `payment_msg_proccessing` consists of two nodes, each with a parallelism of 
1. 

[GitHub] [flink-playgrounds] shuiqiangchen commented on a change in pull request #16: [FLINK-19145][walkthroughs] Add PyFlink-walkthrough to Flink playground.

2020-10-01 Thread GitBox


shuiqiangchen commented on a change in pull request #16:
URL: https://github.com/apache/flink-playgrounds/pull/16#discussion_r498606570



##
File path: pyflink-walkthrough/README.md
##
@@ -0,0 +1,140 @@
+# pyflink-walkthrough
+
+## Background
+
+In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
+
+* Reading data from a Kafka source;
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Performing a simple aggregation over the source data;
+* Writing the results to Elasticsearch and visualizing them in Kibana.
+
+The environment is based on Docker Compose, so the only requirement is that 
you have [Docker](https://docs.docker.com/get-docker/) 
+installed in your machine.
+
+### Kafka
+You will be using Kafka to store sample input data about payment transactions. 
A simple data generator 
[generate_source_data.py](generator/generate_source_data.py) is provided to
+continuously write new records to the `payment_msg` Kafka topic. Each record 
is structured as follows:
+ 
+`{"createTime": "2020-08-12 06:29:02", "orderId": 1597213797, "payAmount": 
28306.44976403719, "payPlatform": 0, "provinceId": 4}`
+
+* `createTime`: The creation time of the transaction. 
+* `orderId`: The id of the current transaction.
+* `payAmount`: The amount being paid with this transaction.
+* `payPlatform`: The platform used to create this payment: pc or mobile.
+* `provinceId`: The id of the province for the user. 
+
+You can use the following command to read data from the Kafka topic and check 
whether it's generated correctly:
+```shell script
+$ docker-compose exec kafka kafka-console-consumer.sh --bootstrap-server 
kafka:9092 --topic payment_msg
+{"createTime":"2020-07-27 
09:25:32.77","orderId":1595841867217,"payAmount":7732.44,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.231","orderId":1595841867218,"payAmount":75774.05,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.72","orderId":1595841867219,"payAmount":65908.55,"payPlatform":0,"provinceId":0}
+{"createTime":"2020-07-27 
09:25:34.216","orderId":1595841867220,"payAmount":15341.11,"payPlatform":0,"provinceId":1}
+{"createTime":"2020-07-27 
09:25:34.698","orderId":1595841867221,"payAmount":37504.42,"payPlatform":0,"provinceId":0}
+```
+You can also create a new topic by executing the following command:
+```shell script
+$ docker-compose exec kafka kafka-topics.sh --bootstrap-server kafka:9092 
--create --topic  --partitions 8 --replication-factor 1
+```
+
+### PyFlink
+
+The transaction data will be processed with PyFlink using the Python script 
[payment_msg_processing.py](payment_msg_proccessing.py).
+This script will first map the `provinceId` in the input records to its 
corresponding province name using a Python UDF, 
+and then compute the sum of the transaction amounts for each province.
+
+### ElasticSearch
+
+ElasticSearch is used to store the results and to provide an efficient query 
service.
+
+### Kibana
+
+Kibana is an open source data visualization dashboard for ElasticSearch. You 
will use it to visualize 
+the results of your PyFlink pipeline.
+
+## Setup
+
+As mentioned, the environment for this walkthrough is based on Docker Compose; 
It uses a custom image
+to spin up Flink (JobManager+TaskManager), Kafka+Zookeeper, the data 
generator, and Elasticsearch+Kibana containers.
+
+You can find the [docker-compose.yaml](docker-compose.yml) file of the 
pyflink-walkthrough in the `pyflink-walkthrough` root directory.
+
+### Building the Docker image
+
+First, build the Docker image by running:
+
+```bash
+docker-compose build
+```
+
+### Starting the Playground
+
+Once the Docker image build is complete, run the following command to start 
the playground:
+
+```bash
+docker-compose up -d
+```
+
+One way of checking if the playground was successfully started is to access 
some of the services that are exposed:
+
+1. visiting Flink Web UI [http://localhost:8081](http://localhost:8081).
+2. visiting Elasticsearch [http://localhost:9200](http://localhost:9200).

Review comment:
   This simply indicates that the Elasticsearch service is successfully up  
when you visit this link. 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-playgrounds] shuiqiangchen commented on a change in pull request #16: [FLINK-19145][walkthroughs] Add PyFlink-walkthrough to Flink playground.

2020-10-01 Thread GitBox


shuiqiangchen commented on a change in pull request #16:
URL: https://github.com/apache/flink-playgrounds/pull/16#discussion_r498605245



##
File path: pyflink-walkthrough/README.md
##
@@ -0,0 +1,140 @@
+# pyflink-walkthrough
+
+## Background
+
+In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
+
+* Reading data from a Kafka source;
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Performing a simple aggregation over the source data;
+* Writing the results to Elasticsearch and visualizing them in Kibana.
+
+The environment is based on Docker Compose, so the only requirement is that 
you have [Docker](https://docs.docker.com/get-docker/) 
+installed in your machine.
+
+### Kafka
+You will be using Kafka to store sample input data about payment transactions. 
A simple data generator 
[generate_source_data.py](generator/generate_source_data.py) is provided to
+continuously write new records to the `payment_msg` Kafka topic. Each record 
is structured as follows:
+ 
+`{"createTime": "2020-08-12 06:29:02", "orderId": 1597213797, "payAmount": 
28306.44976403719, "payPlatform": 0, "provinceId": 4}`
+
+* `createTime`: The creation time of the transaction. 
+* `orderId`: The id of the current transaction.
+* `payAmount`: The amount being paid with this transaction.
+* `payPlatform`: The platform used to create this payment: pc or mobile.
+* `provinceId`: The id of the province for the user. 
+
+You can use the following command to read data from the Kafka topic and check 
whether it's generated correctly:
+```shell script
+$ docker-compose exec kafka kafka-console-consumer.sh --bootstrap-server 
kafka:9092 --topic payment_msg
+{"createTime":"2020-07-27 
09:25:32.77","orderId":1595841867217,"payAmount":7732.44,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.231","orderId":1595841867218,"payAmount":75774.05,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.72","orderId":1595841867219,"payAmount":65908.55,"payPlatform":0,"provinceId":0}
+{"createTime":"2020-07-27 
09:25:34.216","orderId":1595841867220,"payAmount":15341.11,"payPlatform":0,"provinceId":1}
+{"createTime":"2020-07-27 
09:25:34.698","orderId":1595841867221,"payAmount":37504.42,"payPlatform":0,"provinceId":0}
+```
+You can also create a new topic by executing the following command:
+```shell script
+$ docker-compose exec kafka kafka-topics.sh --bootstrap-server kafka:9092 
--create --topic  --partitions 8 --replication-factor 1
+```
+
+### PyFlink
+
+The transaction data will be processed with PyFlink using the Python script 
[payment_msg_processing.py](payment_msg_proccessing.py).
+This script will first map the `provinceId` in the input records to its 
corresponding province name using a Python UDF, 
+and then compute the sum of the transaction amounts for each province.
+
+### ElasticSearch
+
+ElasticSearch is used to store the results and to provide an efficient query 
service.
+
+### Kibana
+
+Kibana is an open source data visualization dashboard for ElasticSearch. You 
will use it to visualize 
+the results of your PyFlink pipeline.
+
+## Setup
+
+As mentioned, the environment for this walkthrough is based on Docker Compose; 
It uses a custom image
+to spin up Flink (JobManager+TaskManager), Kafka+Zookeeper, the data 
generator, and Elasticsearch+Kibana containers.
+
+You can find the [docker-compose.yaml](docker-compose.yml) file of the 
pyflink-walkthrough in the `pyflink-walkthrough` root directory.
+
+### Building the Docker image
+
+First, build the Docker image by running:
+
+```bash
+docker-compose build
+```
+
+### Starting the Playground
+
+Once the Docker image build is complete, run the following command to start 
the playground:
+
+```bash
+docker-compose up -d
+```
+
+One way of checking if the playground was successfully started is to access 
some of the services that are exposed:
+
+1. visiting Flink Web UI [http://localhost:8081](http://localhost:8081).
+2. visiting Elasticsearch [http://localhost:9200](http://localhost:9200).
+3. visiting Kibana [http://localhost:5601](http://localhost:5601).
+
+**Note:** you may need to wait around 1 minute before all the services come up.
+
+### Stopping the Playground
+
+To stop the playground, run the following command:
+
+```bash
+docker-compose down
+```

Review comment:
   Yes, it should be placed in the end of the walkthrough.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13375: [FLINK-18815][filesystems][tests] Join SafetyNetCloseableRegistry's reaper thread before assertion

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13375:
URL: https://github.com/apache/flink/pull/13375#issuecomment-691410175


   
   ## CI report:
   
   * 0e3916e68d38b432ebc33248d3a1ce7f7213b1f8 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=6968)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=6572)
 
   * ce25c6bbc562da022b86aaa99d24f40c1d159b41 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7160)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19489) SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets stuck

2020-10-01 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-19489:

Labels: test-stability  (was: )

> SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets stuck
> ---
>
> Key: FLINK-19489
> URL: https://issues.apache.org/jira/browse/FLINK-19489
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Common
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=298e20ef-7951-5965-0e79-ea664ddc435e=b4cd3436-dbe8-556d-3bca-42f92c3cbf2f
> {code}
> 020-10-01T21:55:34.9982203Z "main" #1 prio=5 os_prio=0 cpu=1048.80ms 
> elapsed=921.99s tid=0x7f8c00015800 nid=0xf6e in Object.wait()  
> [0x7f8c06648000]
> 2020-10-01T21:55:34.9982807Zjava.lang.Thread.State: WAITING (on object 
> monitor)
> 2020-10-01T21:55:34.9983177Z  at 
> java.lang.Object.wait(java.base@11.0.7/Native Method)
> 2020-10-01T21:55:34.9983871Z  - waiting on <0x8e0be190> (a 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
> 2020-10-01T21:55:34.9984581Z  at 
> java.lang.Thread.join(java.base@11.0.7/Thread.java:1305)
> 2020-10-01T21:55:34.9985433Z  - waiting to re-lock in wait() 
> <0x8e0be190> (a 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
> 2020-10-01T21:55:34.9985998Z  at 
> org.apache.flink.core.testutils.CheckedThread.trySync(CheckedThread.java:112)
> 2020-10-01T21:55:34.9986511Z  at 
> org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:100)
> 2020-10-01T21:55:34.9987004Z  at 
> org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:89)
> 2020-10-01T21:55:34.9987707Z  at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread.shutdown(SplitFetcherTest.java:301)
> 2020-10-01T21:55:34.9988427Z  at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent(SplitFetcherTest.java:131)
> 2020-10-01T21:55:34.9989025Z  at 
> jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(java.base@11.0.7/Native 
> Method)
> 2020-10-01T21:55:34.9989531Z  at 
> jdk.internal.reflect.NativeMethodAccessorImpl.invoke(java.base@11.0.7/NativeMethodAccessorImpl.java:62)
> 2020-10-01T21:55:34.9990117Z  at 
> jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(java.base@11.0.7/DelegatingMethodAccessorImpl.java:43)
> 2020-10-01T21:55:34.9990626Z  at 
> java.lang.reflect.Method.invoke(java.base@11.0.7/Method.java:566)
> 2020-10-01T21:55:34.9991078Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-10-01T21:55:34.9991602Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-10-01T21:55:34.9992119Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-10-01T21:55:34.9992749Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-10-01T21:55:34.9993229Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-10-01T21:55:34.9993700Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-10-01T21:55:34.9994202Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-10-01T21:55:34.9994670Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-10-01T21:55:34.9995098Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-10-01T21:55:34.9995524Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-10-01T21:55:34.9995965Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-10-01T21:55:34.9996403Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-10-01T21:55:34.9996816Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-10-01T21:55:34.9997268Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-10-01T21:55:34.9997695Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-10-01T21:55:34.9998077Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-10-01T21:55:34.9998510Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-10-01T21:55:34.9998941Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-10-01T21:55:34.380Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-10-01T21:55:34.815Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 

[jira] [Created] (FLINK-19489) SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets stuck

2020-10-01 Thread Dian Fu (Jira)
Dian Fu created FLINK-19489:
---

 Summary: SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets 
stuck
 Key: FLINK-19489
 URL: https://issues.apache.org/jira/browse/FLINK-19489
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Common
Affects Versions: 1.12.0
Reporter: Dian Fu


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=298e20ef-7951-5965-0e79-ea664ddc435e=b4cd3436-dbe8-556d-3bca-42f92c3cbf2f

{code}
020-10-01T21:55:34.9982203Z "main" #1 prio=5 os_prio=0 cpu=1048.80ms 
elapsed=921.99s tid=0x7f8c00015800 nid=0xf6e in Object.wait()  
[0x7f8c06648000]
2020-10-01T21:55:34.9982807Zjava.lang.Thread.State: WAITING (on object 
monitor)
2020-10-01T21:55:34.9983177Zat 
java.lang.Object.wait(java.base@11.0.7/Native Method)
2020-10-01T21:55:34.9983871Z- waiting on <0x8e0be190> (a 
org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
2020-10-01T21:55:34.9984581Zat 
java.lang.Thread.join(java.base@11.0.7/Thread.java:1305)
2020-10-01T21:55:34.9985433Z- waiting to re-lock in wait() 
<0x8e0be190> (a 
org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
2020-10-01T21:55:34.9985998Zat 
org.apache.flink.core.testutils.CheckedThread.trySync(CheckedThread.java:112)
2020-10-01T21:55:34.9986511Zat 
org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:100)
2020-10-01T21:55:34.9987004Zat 
org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:89)
2020-10-01T21:55:34.9987707Zat 
org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread.shutdown(SplitFetcherTest.java:301)
2020-10-01T21:55:34.9988427Zat 
org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent(SplitFetcherTest.java:131)
2020-10-01T21:55:34.9989025Zat 
jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(java.base@11.0.7/Native 
Method)
2020-10-01T21:55:34.9989531Zat 
jdk.internal.reflect.NativeMethodAccessorImpl.invoke(java.base@11.0.7/NativeMethodAccessorImpl.java:62)
2020-10-01T21:55:34.9990117Zat 
jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(java.base@11.0.7/DelegatingMethodAccessorImpl.java:43)
2020-10-01T21:55:34.9990626Zat 
java.lang.reflect.Method.invoke(java.base@11.0.7/Method.java:566)
2020-10-01T21:55:34.9991078Zat 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
2020-10-01T21:55:34.9991602Zat 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
2020-10-01T21:55:34.9992119Zat 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
2020-10-01T21:55:34.9992749Zat 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
2020-10-01T21:55:34.9993229Zat 
org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
2020-10-01T21:55:34.9993700Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
2020-10-01T21:55:34.9994202Zat 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
2020-10-01T21:55:34.9994670Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-10-01T21:55:34.9995098Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-10-01T21:55:34.9995524Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-10-01T21:55:34.9995965Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-10-01T21:55:34.9996403Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-10-01T21:55:34.9996816Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-10-01T21:55:34.9997268Zat 
org.junit.runners.Suite.runChild(Suite.java:128)
2020-10-01T21:55:34.9997695Zat 
org.junit.runners.Suite.runChild(Suite.java:27)
2020-10-01T21:55:34.9998077Zat 
org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
2020-10-01T21:55:34.9998510Zat 
org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
2020-10-01T21:55:34.9998941Zat 
org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
2020-10-01T21:55:34.380Zat 
org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
2020-10-01T21:55:34.815Zat 
org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
2020-10-01T21:55:35.226Zat 
org.junit.runners.ParentRunner.run(ParentRunner.java:363)
2020-10-01T21:55:35.662Zat 
org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
2020-10-01T21:55:35.0001190Zat 
org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
2020-10-01T21:55:35.0001746Zat 

[jira] [Updated] (FLINK-19489) SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets stuck

2020-10-01 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-19489:

Issue Type: Bug  (was: Improvement)

> SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent gets stuck
> ---
>
> Key: FLINK-19489
> URL: https://issues.apache.org/jira/browse/FLINK-19489
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=298e20ef-7951-5965-0e79-ea664ddc435e=b4cd3436-dbe8-556d-3bca-42f92c3cbf2f
> {code}
> 020-10-01T21:55:34.9982203Z "main" #1 prio=5 os_prio=0 cpu=1048.80ms 
> elapsed=921.99s tid=0x7f8c00015800 nid=0xf6e in Object.wait()  
> [0x7f8c06648000]
> 2020-10-01T21:55:34.9982807Zjava.lang.Thread.State: WAITING (on object 
> monitor)
> 2020-10-01T21:55:34.9983177Z  at 
> java.lang.Object.wait(java.base@11.0.7/Native Method)
> 2020-10-01T21:55:34.9983871Z  - waiting on <0x8e0be190> (a 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
> 2020-10-01T21:55:34.9984581Z  at 
> java.lang.Thread.join(java.base@11.0.7/Thread.java:1305)
> 2020-10-01T21:55:34.9985433Z  - waiting to re-lock in wait() 
> <0x8e0be190> (a 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread)
> 2020-10-01T21:55:34.9985998Z  at 
> org.apache.flink.core.testutils.CheckedThread.trySync(CheckedThread.java:112)
> 2020-10-01T21:55:34.9986511Z  at 
> org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:100)
> 2020-10-01T21:55:34.9987004Z  at 
> org.apache.flink.core.testutils.CheckedThread.sync(CheckedThread.java:89)
> 2020-10-01T21:55:34.9987707Z  at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest$QueueDrainerThread.shutdown(SplitFetcherTest.java:301)
> 2020-10-01T21:55:34.9988427Z  at 
> org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherTest.testNotifiesWhenGoingIdleConcurrent(SplitFetcherTest.java:131)
> 2020-10-01T21:55:34.9989025Z  at 
> jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(java.base@11.0.7/Native 
> Method)
> 2020-10-01T21:55:34.9989531Z  at 
> jdk.internal.reflect.NativeMethodAccessorImpl.invoke(java.base@11.0.7/NativeMethodAccessorImpl.java:62)
> 2020-10-01T21:55:34.9990117Z  at 
> jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(java.base@11.0.7/DelegatingMethodAccessorImpl.java:43)
> 2020-10-01T21:55:34.9990626Z  at 
> java.lang.reflect.Method.invoke(java.base@11.0.7/Method.java:566)
> 2020-10-01T21:55:34.9991078Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-10-01T21:55:34.9991602Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-10-01T21:55:34.9992119Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-10-01T21:55:34.9992749Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-10-01T21:55:34.9993229Z  at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-10-01T21:55:34.9993700Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-10-01T21:55:34.9994202Z  at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-10-01T21:55:34.9994670Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-10-01T21:55:34.9995098Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-10-01T21:55:34.9995524Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-10-01T21:55:34.9995965Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-10-01T21:55:34.9996403Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-10-01T21:55:34.9996816Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-10-01T21:55:34.9997268Z  at 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-10-01T21:55:34.9997695Z  at 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-10-01T21:55:34.9998077Z  at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-10-01T21:55:34.9998510Z  at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-10-01T21:55:34.9998941Z  at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-10-01T21:55:34.380Z  at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-10-01T21:55:34.815Z  at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-10-01T21:55:35.226Z  

[jira] [Updated] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-19027:

Affects Version/s: 1.12.0

> UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel
>  failed because of test timeout
> 
>
> Key: FLINK-19027
> URL: https://issues.apache.org/jira/browse/FLINK-19027
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.12.0, 1.11.2
>Reporter: Dian Fu
>Priority: Major
>  Labels: test-stability
> Fix For: 1.12.0, 1.11.3
>
>
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=5789=logs=119bbba7-f5e3-5e08-e72d-09f1529665de=ec103906-d047-5b8a-680e-05fc000dfca9]
> {code}
> 2020-08-22T21:13:05.5315459Z [ERROR] 
> shouldPerformUnalignedCheckpointOnParallelRemoteChannel(org.apache.flink.test.checkpointing.UnalignedCheckpointITCase)
>   Time elapsed: 300.075 s  <<< ERROR!
> 2020-08-22T21:13:05.5316451Z org.junit.runners.model.TestTimedOutException: 
> test timed out after 300 seconds
> 2020-08-22T21:13:05.5317432Z  at sun.misc.Unsafe.park(Native Method)
> 2020-08-22T21:13:05.5317799Z  at 
> java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
> 2020-08-22T21:13:05.5318247Z  at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
> 2020-08-22T21:13:05.5318885Z  at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
> 2020-08-22T21:13:05.5327035Z  at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
> 2020-08-22T21:13:05.5328114Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
> 2020-08-22T21:13:05.5328869Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1719)
> 2020-08-22T21:13:05.5329482Z  at 
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:74)
> 2020-08-22T21:13:05.5330138Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1699)
> 2020-08-22T21:13:05.5330771Z  at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:1681)
> 2020-08-22T21:13:05.5331351Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.execute(UnalignedCheckpointITCase.java:158)
> 2020-08-22T21:13:05.5332015Z  at 
> org.apache.flink.test.checkpointing.UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel(UnalignedCheckpointITCase.java:140)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205904#comment-17205904
 ] 

Dian Fu edited comment on FLINK-19027 at 10/2/20, 12:43 AM:


Another instance on master:
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=ae4f8708-9994-57d3-c2d7-b892156e7812=5c8e7682-d68f-54d1-16a2-a09310218a49

Found the following exceptions in the log, hope it could help to identify the 
issue:
{code}
21:38:30,457 [   Map (1/1)] WARN  
org.apache.flink.runtime.taskmanager.Task[] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:164)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:179)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:152)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:67)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:368)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:185)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:594)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:558) 
~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:722) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:547) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_242]
21:38:30,458 [   Map (1/1)] INFO  
org.apache.flink.runtime.taskmanager.Task[] - Freeing task 
resources for Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0).
21:38:30,458 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor   [] - 
Un-registering task and sending final execution state FAILED to JobManager for 
task Map (1/1) 
cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0.
21:38:30,459 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph   [] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED on 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@2535c61e.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]

[jira] [Comment Edited] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205904#comment-17205904
 ] 

Dian Fu edited comment on FLINK-19027 at 10/2/20, 12:38 AM:


Another instance:
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=ae4f8708-9994-57d3-c2d7-b892156e7812=5c8e7682-d68f-54d1-16a2-a09310218a49

Found the following exceptions in the log, hope it could help to identify the 
issue:
{code}
21:38:30,457 [   Map (1/1)] WARN  
org.apache.flink.runtime.taskmanager.Task[] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:164)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:179)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:152)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:67)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:368)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:185)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:594)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:558) 
~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:722) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:547) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_242]
21:38:30,458 [   Map (1/1)] INFO  
org.apache.flink.runtime.taskmanager.Task[] - Freeing task 
resources for Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0).
21:38:30,458 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor   [] - 
Un-registering task and sending final execution state FAILED to JobManager for 
task Map (1/1) 
cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0.
21:38:30,459 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph   [] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED on 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@2535c61e.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 

[jira] [Commented] (FLINK-19027) UnalignedCheckpointITCase.shouldPerformUnalignedCheckpointOnParallelRemoteChannel failed because of test timeout

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205904#comment-17205904
 ] 

Dian Fu commented on FLINK-19027:
-

Another instance:
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7158=logs=ae4f8708-9994-57d3-c2d7-b892156e7812=5c8e7682-d68f-54d1-16a2-a09310218a49

{code}
21:38:30,457 [   Map (1/1)] WARN  
org.apache.flink.runtime.taskmanager.Task[] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:164)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.processElement(StreamTaskNetworkInput.java:179)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamTaskNetworkInput.emitNext(StreamTaskNetworkInput.java:152)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:67)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:368)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:185)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:594)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:558) 
~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:722) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at org.apache.flink.runtime.taskmanager.Task.run(Task.java:547) 
[flink-runtime_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_242]
21:38:30,458 [   Map (1/1)] INFO  
org.apache.flink.runtime.taskmanager.Task[] - Freeing task 
resources for Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0).
21:38:30,458 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.taskexecutor.TaskExecutor   [] - 
Un-registering task and sending final execution state FAILED to JobManager for 
task Map (1/1) 
cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0.
21:38:30,459 [flink-akka.actor.default-dispatcher-4] INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph   [] - Map (1/1) 
(cb2b6ee8f94a6516932dfa923bfede3f_0a448493b4782967b150582570326227_0_0) 
switched from RUNNING to FAILED on 
org.apache.flink.runtime.jobmaster.slotpool.SingleLogicalSlot@2535c61e.
java.lang.Exception: Failing map @ 4 (0 attempt); last value 391
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.failMapper(UnalignedCheckpointITCase.java:616)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.checkFail(UnalignedCheckpointITCase.java:611)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:605)
 ~[test-classes/:?]
at 
org.apache.flink.test.checkpointing.UnalignedCheckpointITCase$FailingMapper.map(UnalignedCheckpointITCase.java:580)
 ~[test-classes/:?]
at 
org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:41)
 ~[flink-streaming-java_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
at 
org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:164)
 

[GitHub] [flink] flinkbot edited a comment on pull request #13375: [FLINK-18815][filesystems][tests] Join SafetyNetCloseableRegistry's reaper thread before assertion

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13375:
URL: https://github.com/apache/flink/pull/13375#issuecomment-691410175


   
   ## CI report:
   
   * 0e3916e68d38b432ebc33248d3a1ce7f7213b1f8 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=6968)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=6572)
 
   * ce25c6bbc562da022b86aaa99d24f40c1d159b41 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18366) Track E2E test durations centrally

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18366?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205902#comment-17205902
 ] 

Dian Fu commented on FLINK-18366:
-

slow run: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7153=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529

> Track E2E test durations centrally
> --
>
> Key: FLINK-18366
> URL: https://issues.apache.org/jira/browse/FLINK-18366
> Project: Flink
>  Issue Type: Improvement
>  Components: Build System / CI
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Major
>
> Every now and then, our E2E tests start timing out (see FLINK-16795), because 
> they hit the currently configured time-limit.
> To better understand what the expected E2E time, and potential performance 
> regressions, we should track the test execution duration centrally.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] kezhuw commented on a change in pull request #13375: [FLINK-18815][filesystems][tests] Join SafetyNetCloseableRegistry's reaper thread before assertion

2020-10-01 Thread GitBox


kezhuw commented on a change in pull request #13375:
URL: https://github.com/apache/flink/pull/13375#discussion_r498570772



##
File path: 
flink-core/src/main/java/org/apache/flink/core/fs/SafetyNetCloseableRegistry.java
##
@@ -134,10 +135,17 @@ public void close() throws IOException {
--GLOBAL_SAFETY_NET_REGISTRY_COUNT;
if (0 == GLOBAL_SAFETY_NET_REGISTRY_COUNT) {
REAPER_THREAD.interrupt();

Review comment:
   All reclamation related classes and methods, say, 
`CloseableReaperThread`, `PhantomDelegatingCloseableRef` and 
`SafetyNetCloseableRegistry(Supplier 
reaperThreadSupplier)`, are private or visible for testing, normal clients 
should care only about the reclamation of leaking resources guarantee exported 
by `SafetyNetCloseableRegistry`, "join on the reaper thread" does not 
contribute to that exported guarantee.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] kezhuw commented on a change in pull request #13375: [FLINK-18815][filesystems][tests] Join SafetyNetCloseableRegistry's reaper thread before assertion

2020-10-01 Thread GitBox


kezhuw commented on a change in pull request #13375:
URL: https://github.com/apache/flink/pull/13375#discussion_r498570772



##
File path: 
flink-core/src/main/java/org/apache/flink/core/fs/SafetyNetCloseableRegistry.java
##
@@ -134,10 +135,17 @@ public void close() throws IOException {
--GLOBAL_SAFETY_NET_REGISTRY_COUNT;
if (0 == GLOBAL_SAFETY_NET_REGISTRY_COUNT) {
REAPER_THREAD.interrupt();

Review comment:
   All reclamation related classes and methods, say, 
`CloseableReaperThread`, `PhantomDelegatingCloseableRef` and 
`SafetyNetCloseableRegistry(Supplier 
reaperThreadSupplier)`, are private or visible for testing, normal clients 
should care only about the reclamation of leaking resources guarantee exported 
by `SafetyNetCloseableRegistry`, "join on the reaper thread" does not belong to 
that exported guarantee.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13040: [FLINK-17073] [checkpointing] checkpointing backpressure if there are too many checkpoints to clean

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13040:
URL: https://github.com/apache/flink/pull/13040#issuecomment-667140037


   
   ## CI report:
   
   * 31abcb25fdf72394dbfa91d0ba9e151968375ad7 UNKNOWN
   * db99952f8166cb4d6337dcb3815e72ff2b49604b UNKNOWN
   * 33af84f70ab92a76818f1a4392fecb0c4a8e62f9 UNKNOWN
   * 12d5698e9185b3b97e5323485ee72617c4e535f9 UNKNOWN
   * b7162b6109b03f905f2c52cc22002b198bc9438e UNKNOWN
   * d83bf413dd7661c4ee87a692b80147f2a6750f46 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7157)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13529: [FLINK-19473] Implement multi inputs sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13529:
URL: https://github.com/apache/flink/pull/13529#issuecomment-702193043


   
   ## CI report:
   
   * 6270bf09d731ac358ca621e4150c817153882eb9 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7155)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13521:
URL: https://github.com/apache/flink/pull/13521#issuecomment-701282459


   
   ## CI report:
   
   * c5164bceb05c9e11c652fdb2a0cefe289af0ec1b UNKNOWN
   * 39e66d00bae3c21b41c12881cd2680f74ffbf2e1 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7154)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] StephanEwen commented on pull request #13512: [FLINK-19457][core] Add a number sequence generating source for the New Source API.

2020-10-01 Thread GitBox


StephanEwen commented on pull request #13512:
URL: https://github.com/apache/flink/pull/13512#issuecomment-702303125


   No worries, let's figure this out together.
   
   ### About the flink-core package structure.
   
   Agree, `flink-core`'s package structure is wild.
   
   Most of it is because of the pre-streaming origin, and there was not too 
much thought there. `api.common` was shared between Batch Java and Batch Scala 
APIs. `api.java` was originally in a different project (Batch Java API) and 
moved to core when streaming came into the mix to be shared with the Streaming 
Java API.
   
   `core.io` was IIRC IO-related stuff that was not part of the API, but 
indirectly relevant in core API classes, like the utils around the managed 
memory abstraction. `api.common.io` was for API-related I/O, like sources, 
sinks.
   
   A lot of stuff was also added over time in the wrong places, admittedly, 
because few people in the early days paid attention to package structure. The 
packages stuck because we didn't want to break user applications. For Flink 
2.0, we should try and simplify this.
   
   tl;dr - it is the way it is for historic reasons and because the "not 
breaking application" consideration ended up trumping the "clean up packages" 
consideration. Seems the structure was not getting too much in the way of users 
after all.
   
   ### "flink-connector-base" versus "flink-core"
   
   What I would like to keep is that this source implementation stays in 
`flink-core` and does not move to another module. We have a lot of dependencies 
already and it really doesn't hurt in my opinion to have that implementation 
there in the core (given that it is lightweight and has no dependencies).
   
   The Split-Reader API could be moved to flink-core, theoretically. I didn't 
argue for that in the original PR, because it is not used for simple sources 
(like here), but typically for sources with external systems that need separate 
I/O threads. These sources would be in separate modules anyways. So having the 
Split Reader API in a separate module would not make any difference to users, 
so I just went with your suggestion there (if it makes no difference, I 
typically go with what the code author suggests). IT has the slight advantage 
even. of being more flexible with test dependencies, so Integration Tests can 
be in the module directly.
   
   ### Packages in this PR
   
   About the `lib` and `util` package names, I am not opinionated on those. In 
fact, I don't particularly like `lib` either.
   Anything that is reasonably "discoverable" for users would work here from my 
side. Let's try to up with different names or locations, I am definitely open 
to suggestions.
   Maybe we'll move this in a follow-up effort, to help unblock some features 
where I wanted to use this source in the tests.
   
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

root
{{     |-- C0: BIGINT}}
{{     |-- C1: STRING}}
{{     |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 

  was:
I have a table T0 with the following schema -

{{root}}
    |-- C0: BIGINT
    |-- C1: STRING
    |-- blaze_itime: TIMESTAMP(3)

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
{{    |-- C0: BIGINT}}
{{    |-- C1: STRING}}
{{    |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 

  was:
I have a table T0 with the following schema -

{{root}}
   |-- C0: BIGINT
   |-- C1: STRING
   |-- blaze_itime: TIMESTAMP(3)

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
    |-- C0: BIGINT
    |-- C1: STRING
    |-- blaze_itime: TIMESTAMP(3)

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 

  was:
I have a table T0 with the following schema -

{{root}}
{{    |-- C0: BIGINT}}
{{    |-- C1: STRING}}
{{    |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Environment: 
Flink version: 1.11.1

 

  was:
Flink version: 1.11.1

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
{{env.getConfig().enableObjectReuse();}}
{{var settings = EnvironmentSettings.newInstance()}}
{{  .useBlinkPlanner()}}
{{  .inBatchMode()}}
{{  .build();}}


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
>  \{{ |-- C0: BIGINT}}
>  \{{ |-- C1: STRING}}
>  \{{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - INTERVAL '10' MINUTE AND NOW());}}
>  
> {{The generated code for the above query tries to assign a long value to 
> timestamp type and fails to compile with the following exception -}}
>  
> {{java.lang.RuntimeException: Could not instantiate generated class 
> 'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated 
> class 'BatchCalc$14' at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
>  at 
> org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
>  at 
> org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
>  at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
>  at}}
> {{...}}
> {{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
> 21: Assignment conversion not possible from type "long" to type 
> "org.apache.flink.table.data.TimestampData"Caused by: 
> org.codehaus.commons.compiler.CompileException: Line 55, Column 21: 
> Assignment conversion not possible from type "long" to type 
> "org.apache.flink.table.data.TimestampData" at 
> org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
> org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975)
>  at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
> org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}
>  
> The generated code is added as an attachment to the issue.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 

  was:
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
>  \{{ |-- C0: BIGINT}}
>  \{{ |-- C1: STRING}}
>  \{{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - INTERVAL '10' MINUTE AND 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
   |-- C0: BIGINT
   |-- C1: STRING
   |-- blaze_itime: TIMESTAMP(3)

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 

  was:
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
 {{env.getConfig().enableObjectReuse();}}
 {{var settings = EnvironmentSettings.newInstance()}}
 {{  .useBlinkPlanner()}}
 {{  .inBatchMode()}}
 {{  .build();}}

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
>  
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

 

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at}}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

  was:
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at 

...

{{

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

}}

 

The generated code is added as an attachment to the issue.

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
> The Environment has the following configuration parameters -
> {{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
> {{env.getConfig().enableObjectReuse();}}
> {{var settings = EnvironmentSettings.newInstance()}}
> {{  .useBlinkPlanner()}}
> {{  .inBatchMode()}}
> {{  .build();}}
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
>  \{{ |-- C0: BIGINT}}
>  \{{ |-- C1: STRING}}
>  \{{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - INTERVAL '10' 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at 

...

{{

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

}}

 

The generated code is added as an attachment to the issue.

 

  was:
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at ...Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)

 

The generated code is added as an attachment to the issue.

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
> The Environment has the following configuration parameters -
> {{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
> {{env.getConfig().enableObjectReuse();}}
> {{var settings = EnvironmentSettings.newInstance()}}
> {{  .useBlinkPlanner()}}
> {{  .inBatchMode()}}
> {{  .build();}}
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
>  \{{ |-- C0: BIGINT}}
>  \{{ |-- C1: STRING}}
>  \{{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at }}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 

  was:
I have a table T0 with the following schema -

{{root}}
{{ |-- C0: BIGINT}}
{{ |-- C1: STRING}}
{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at }}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

{{Here is the generated code -}}

/* 1 */
/* 2 */ public class BatchCalc$14 extends 
org.apache.flink.table.runtime.operators.TableStreamOperator
/* 3 */ implements 
org.apache.flink.streaming.api.operators.OneInputStreamOperator {
/* 4 */
/* 5 */ private final Object[] references;
/* 6 */ org.apache.flink.table.data.BoxedWrapperRowData out = new 
org.apache.flink.table.data.BoxedWrapperRowData(2);
/* 7 */ private final 
org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
/* 8 */
/* 9 */ public BatchCalc$14(
/* 10 */ Object[] references,
/* 11 */ org.apache.flink.streaming.runtime.tasks.StreamTask task,
/* 12 */ org.apache.flink.streaming.api.graph.StreamConfig config,
/* 13 */ org.apache.flink.streaming.api.operators.Output output,
/* 14 */ org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
processingTimeService) throws Exception {
/* 15 */ this.references = references;
/* 16 */ 
/* 17 */ this.setup(task, config, output);
/* 18 */ if (this instanceof 
org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
/* 19 */ 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Description: 
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at ...Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)

 

The generated code is added as an attachment to the issue.

 

  was:
I have a table T0 with the following schema -

{{root}}
 \{{ |-- C0: BIGINT}}
 \{{ |-- C1: STRING}}
 \{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at }}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

The generated code is added as an attachment to the issue.

 


> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
> The Environment has the following configuration parameters -
> {{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
> {{env.getConfig().enableObjectReuse();}}
> {{var settings = EnvironmentSettings.newInstance()}}
> {{  .useBlinkPlanner()}}
> {{  .inBatchMode()}}
> {{  .build();}}
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
>  \{{ |-- C0: BIGINT}}
>  \{{ |-- C1: STRING}}
>  \{{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - INTERVAL '10' MINUTE AND 

[jira] [Updated] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19488?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Satyam Shekhar updated FLINK-19488:
---
Attachment: code.java

> Failed compilation of generated class
> -
>
> Key: FLINK-19488
> URL: https://issues.apache.org/jira/browse/FLINK-19488
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.11.1
> Environment: Flink version: 1.11.1
> The Environment has the following configuration parameters -
> {{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
> {{env.getConfig().enableObjectReuse();}}
> {{var settings = EnvironmentSettings.newInstance()}}
> {{  .useBlinkPlanner()}}
> {{  .inBatchMode()}}
> {{  .build();}}
>Reporter: Satyam Shekhar
>Priority: Major
> Attachments: code.java
>
>
> I have a table T0 with the following schema -
> {{root}}
> {{ |-- C0: BIGINT}}
> {{ |-- C1: STRING}}
> {{ |-- blaze_itime: TIMESTAMP(3)}}
>  
> The following SQL query fails for the above table - 
> {{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime 
> BETWEEN NOW() - INTERVAL '10' MINUTE AND NOW());}}
>  
> {{The generated code for the above query tries to assign a long value to 
> timestamp type and fails to compile with the following exception -}}
> {{java.lang.RuntimeException: Could not instantiate generated class 
> 'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated 
> class 'BatchCalc$14' at 
> org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
>  at 
> org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
>  at 
> org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
>  at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
>  at }}
> {{...}}
> {{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
> 21: Assignment conversion not possible from type "long" to type 
> "org.apache.flink.table.data.TimestampData"Caused by: 
> org.codehaus.commons.compiler.CompileException: Line 55, Column 21: 
> Assignment conversion not possible from type "long" to type 
> "org.apache.flink.table.data.TimestampData" at 
> org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
> org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975)
>  at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
> org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}
>  
> {{Here is the generated code -}}
> /* 1 */
> /* 2 */ public class BatchCalc$14 extends 
> org.apache.flink.table.runtime.operators.TableStreamOperator
> /* 3 */ implements 
> org.apache.flink.streaming.api.operators.OneInputStreamOperator {
> /* 4 */
> /* 5 */ private final Object[] references;
> /* 6 */ org.apache.flink.table.data.BoxedWrapperRowData out = new 
> org.apache.flink.table.data.BoxedWrapperRowData(2);
> /* 7 */ private final 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
> org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
> /* 8 */
> /* 9 */ public BatchCalc$14(
> /* 10 */ Object[] references,
> /* 11 */ org.apache.flink.streaming.runtime.tasks.StreamTask task,
> /* 12 */ org.apache.flink.streaming.api.graph.StreamConfig config,
> /* 13 */ org.apache.flink.streaming.api.operators.Output output,
> /* 14 */ org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
> processingTimeService) throws Exception {
> /* 15 */ this.references = references;
> /* 16 */ 
> /* 17 */ this.setup(task, config, output);
> /* 18 */ if (this instanceof 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
> /* 19 */ ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) 
> this)
> /* 20 */ .setProcessingTimeService(processingTimeService);
> /* 21 */ }
> /* 22 */ }
> /* 23 */
> /* 24 */ @Override
> /* 25 */ public void open() throws Exception {
> /* 26 */ super.open();
> /* 27 */ 
> /* 28 */ }
> /* 29 */
> /* 30 */ @Override
> /* 31 */ public void 
> processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
> element) throws Exception {
> /* 32 */ org.apache.flink.table.data.RowData in1 = 
> (org.apache.flink.table.data.RowData) element.getValue();
> /* 33 */ 
> /* 34 */ org.apache.flink.table.data.TimestampData field$2;
> /* 35 */ boolean isNull$2;
> /* 36 */ boolean isNull$3;
> /* 37 */ org.apache.flink.table.data.TimestampData result$4;
> /* 38 */ boolean isNull$5;
> /* 39 */ org.apache.flink.table.data.TimestampData result$6;
> /* 40 */ boolean isNull$7;
> /* 41 */ boolean result$8;
> /* 42 */ boolean isNull$9;
> /* 43 */ 

[jira] [Created] (FLINK-19488) Failed compilation of generated class

2020-10-01 Thread Satyam Shekhar (Jira)
Satyam Shekhar created FLINK-19488:
--

 Summary: Failed compilation of generated class
 Key: FLINK-19488
 URL: https://issues.apache.org/jira/browse/FLINK-19488
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Runtime
Affects Versions: 1.11.1
 Environment: Flink version: 1.11.1

The Environment has the following configuration parameters -

{{env.setParallelism(Integer.getInteger("flinkParallelism", 2));}}
{{env.getConfig().enableObjectReuse();}}
{{var settings = EnvironmentSettings.newInstance()}}
{{  .useBlinkPlanner()}}
{{  .inBatchMode()}}
{{  .build();}}
Reporter: Satyam Shekhar


I have a table T0 with the following schema -

{{root}}
{{ |-- C0: BIGINT}}
{{ |-- C1: STRING}}
{{ |-- blaze_itime: TIMESTAMP(3)}}

 

The following SQL query fails for the above table - 

{{SELECT A.C0 AS output, A.C1 AS dim0 FROM T0 AS A WHERE (A.blaze_itime BETWEEN 
NOW() - INTERVAL '10' MINUTE AND NOW());}}

 

{{The generated code for the above query tries to assign a long value to 
timestamp type and fails to compile with the following exception -}}

{{java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14'java.lang.RuntimeException: Could not instantiate generated class 
'BatchCalc$14' at 
org.apache.flink.table.runtime.generated.GeneratedClass.newInstance(GeneratedClass.java:67)
 at 
org.apache.flink.table.runtime.operators.CodeGenOperatorFactory.createStreamOperator(CodeGenOperatorFactory.java:40)
 at 
org.apache.flink.streaming.api.operators.StreamOperatorFactoryUtil.createOperator(StreamOperatorFactoryUtil.java:70)
 at 
org.apache.flink.streaming.runtime.tasks.OperatorChain.createChainedOperator(OperatorChain.java:470)
 at }}

{{...}}

{{Caused by: org.codehaus.commons.compiler.CompileException: Line 55, Column 
21: Assignment conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData"Caused by: 
org.codehaus.commons.compiler.CompileException: Line 55, Column 21: Assignment 
conversion not possible from type "long" to type 
"org.apache.flink.table.data.TimestampData" at 
org.codehaus.janino.UnitCompiler.compileError(UnitCompiler.java:12124) at 
org.codehaus.janino.UnitCompiler.assignmentConversion(UnitCompiler.java:10975) 
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:3788) at 
org.codehaus.janino.UnitCompiler.access$6100(UnitCompiler.java:215)}}

 

{{Here is the generated code -}}

/* 1 */
/* 2 */ public class BatchCalc$14 extends 
org.apache.flink.table.runtime.operators.TableStreamOperator
/* 3 */ implements 
org.apache.flink.streaming.api.operators.OneInputStreamOperator {
/* 4 */
/* 5 */ private final Object[] references;
/* 6 */ org.apache.flink.table.data.BoxedWrapperRowData out = new 
org.apache.flink.table.data.BoxedWrapperRowData(2);
/* 7 */ private final 
org.apache.flink.streaming.runtime.streamrecord.StreamRecord outElement = new 
org.apache.flink.streaming.runtime.streamrecord.StreamRecord(null);
/* 8 */
/* 9 */ public BatchCalc$14(
/* 10 */ Object[] references,
/* 11 */ org.apache.flink.streaming.runtime.tasks.StreamTask task,
/* 12 */ org.apache.flink.streaming.api.graph.StreamConfig config,
/* 13 */ org.apache.flink.streaming.api.operators.Output output,
/* 14 */ org.apache.flink.streaming.runtime.tasks.ProcessingTimeService 
processingTimeService) throws Exception {
/* 15 */ this.references = references;
/* 16 */ 
/* 17 */ this.setup(task, config, output);
/* 18 */ if (this instanceof 
org.apache.flink.streaming.api.operators.AbstractStreamOperator) {
/* 19 */ ((org.apache.flink.streaming.api.operators.AbstractStreamOperator) 
this)
/* 20 */ .setProcessingTimeService(processingTimeService);
/* 21 */ }
/* 22 */ }
/* 23 */
/* 24 */ @Override
/* 25 */ public void open() throws Exception {
/* 26 */ super.open();
/* 27 */ 
/* 28 */ }
/* 29 */
/* 30 */ @Override
/* 31 */ public void 
processElement(org.apache.flink.streaming.runtime.streamrecord.StreamRecord 
element) throws Exception {
/* 32 */ org.apache.flink.table.data.RowData in1 = 
(org.apache.flink.table.data.RowData) element.getValue();
/* 33 */ 
/* 34 */ org.apache.flink.table.data.TimestampData field$2;
/* 35 */ boolean isNull$2;
/* 36 */ boolean isNull$3;
/* 37 */ org.apache.flink.table.data.TimestampData result$4;
/* 38 */ boolean isNull$5;
/* 39 */ org.apache.flink.table.data.TimestampData result$6;
/* 40 */ boolean isNull$7;
/* 41 */ boolean result$8;
/* 42 */ boolean isNull$9;
/* 43 */ org.apache.flink.table.data.TimestampData result$10;
/* 44 */ boolean isNull$11;
/* 45 */ boolean result$12;
/* 46 */ 
/* 47 */ 
/* 48 */ isNull$2 = in1.isNullAt(2);
/* 49 */ field$2 = null;
/* 50 */ if (!isNull$2) {
/* 51 */ field$2 = in1.getTimestamp(2, 3);
/* 52 */ }
/* 53 */ isNull$3 = false;
/* 54 */ 
/* 55 */ result$4 = 
/* 56 */ org.apache.flink.table.runtime.functions.SqlDateTimeUtils.now()
/* 57 */ ;
/* 58 */ 
/* 59 */ 
/* 60 */ 

[GitHub] [flink] azagrebin commented on a change in pull request #13458: FLINK-18851 [runtime] Add checkpoint type to checkpoint history entries in Web UI

2020-10-01 Thread GitBox


azagrebin commented on a change in pull request #13458:
URL: https://github.com/apache/flink/pull/13458#discussion_r498390360



##
File path: flink-runtime-web/src/test/resources/rest_api_v1.snapshot
##
@@ -557,7 +557,7 @@
   },
   "status" : {
 "type" : "string",
-"enum" : [ "CREATED", "RUNNING", "FAILING", "FAILED", 
"CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", "RECONCILING" ]
+"enum" : [ "INITIALIZING", "CREATED", "RUNNING", "FAILING", 
"FAILED", "CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", 
"RECONCILING" ]

Review comment:
   is this related to this PR?
   also the next similar change in this file?
   
   it does not look related to checkpointing, then it should be addressed at 
least in a separate hotfix commit or better in another PR/Jira issue.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #13458: FLINK-18851 [runtime] Add checkpoint type to checkpoint history entries in Web UI

2020-10-01 Thread GitBox


azagrebin commented on a change in pull request #13458:
URL: https://github.com/apache/flink/pull/13458#discussion_r498392633



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/rest/messages/checkpoints/CheckpointStatistics.java
##
@@ -229,8 +242,8 @@ public static CheckpointStatistics 
generateCheckpointStatistics(AbstractCheckpoi
checkpointStatisticsPerTask.put(
taskStateStat.getJobVertexId(),
new TaskCheckpointStatistics(
-   
checkpointStats.getCheckpointId(),
-   
checkpointStats.getStatus(),
+   
checkpointStats.getCheckpointId(),
+   checkpointStats.getStatus(),

Review comment:
   this change also looks unrelated.
   it looks like a minor code style fix which we usually do not do to preserve 
git history.
   
   if we do the code style fixes then it is usually either
   - massively in files which have a lot of code style violation in separate 
commit
   - or close to related changes

##
File path: flink-runtime-web/src/test/resources/rest_api_v1.snapshot
##
@@ -557,7 +557,7 @@
   },
   "status" : {
 "type" : "string",
-"enum" : [ "CREATED", "RUNNING", "FAILING", "FAILED", 
"CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", "RECONCILING" ]
+"enum" : [ "INITIALIZING", "CREATED", "RUNNING", "FAILING", 
"FAILED", "CANCELLING", "CANCELED", "FINISHED", "RESTARTING", "SUSPENDED", 
"RECONCILING" ]

Review comment:
   is this related to this PR?
   also the next similar change in this file?
   
   it does not look related to checkpointing, then it should addressed at least 
in a separate hotfix commit or better in another PR/Jira issue.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on a change in pull request #13458: FLINK-18851 [runtime] Add checkpoint type to checkpoint history entries in Web UI

2020-10-01 Thread GitBox


azagrebin commented on a change in pull request #13458:
URL: https://github.com/apache/flink/pull/13458#discussion_r498389552



##
File path: flink-runtime-web/src/test/resources/rest_api_v1.snapshot
##
@@ -1254,6 +1263,9 @@
 "num_acknowledged_subtasks" : {
   "type" : "integer"
 },
+"check_point_type" : {

Review comment:
   It should be probably documented in `flink-runtime-web/README.md` if not 
yet anywhere else
   what do you think @zentol ?
   @gm7y8 if it is not documented anywhere else, you can create a JIRA issue 
for that and then open PR.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] becketqin commented on pull request #13512: [FLINK-19457][core] Add a number sequence generating source for the New Source API.

2020-10-01 Thread GitBox


becketqin commented on pull request #13512:
URL: https://github.com/apache/flink/pull/13512#issuecomment-702245190


   @StephanEwen Thanks for the explanation. I checked the classes and structure 
in `flink-core` again, I agree that this PR is in general following the 
existing convention. And I am fine with merge the PR as is. 
   
   That said, I had a hard time to fully understand the structure of 
`flink-core`. It would be great if you can help me understand this a bit more. 
I am wondering if the users would have the same questions like below.
   
   - I was initially expecting the function interfaces in 
`org.apache.api.java.functions` path, but there are only `KeySelector` and 
`NullByteKeySelector`. The actual function interfaces are in 
`org.apache.flink.common.functions` package. What is the difference between 
those two packages?
   - Also, I was not quite sure about the differences between 
`org.apache.flink.api.common.io` and `org.apache.flink.core.io`. 
   - In the existing package paths in `flink-core`, the implementation and 
interface classes are put in the same package. This PR introduces a new package 
`lib`. Should the other implementations also be put in a `lib` path?
   - In the existing package paths, `util` packages usually contains internal 
helper classes for the (abstract) implementations of the interfaces. This PR 
seems putting the abstract implementations of the `SourceReader` / `Enumerator` 
interfaces in the `util` package, and expecting other Iterator based sources to 
leverage that? Or are those classes also internal?
   - The follow up question to the above one, it seems that the 
`IteratorSourceReader`/ `IteratorSourceEnumerator` classes are just another 
base implementation of the `Source` interface, why the base implementation of 
the source in `flink-connector-base` was not put into 
`org.apache.flink.api.connector.source.lib.util`? 
   
   Sorry for sticking to discussion, my main concern is that If we ever want to 
revisit the API and package structure, right now might be the best timing, 
because we are essentially introducing new `Source` APIs and will probably 
deprecate old `SourceFunction` and `InputFormat` APIs in some coming release. 
This would be a natural migration towards new APIs if we want to do that. 
Keeping the API / package structure as is and changing them in the future could 
be even more disruptive.
   
   Thanks again.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] becketqin edited a comment on pull request #13512: [FLINK-19457][core] Add a number sequence generating source for the New Source API.

2020-10-01 Thread GitBox


becketqin edited a comment on pull request #13512:
URL: https://github.com/apache/flink/pull/13512#issuecomment-702245190


   @StephanEwen Thanks for the explanation. I checked the classes and structure 
in `flink-core` again, I agree that this PR is in general following the 
existing convention. And I am fine with merging the PR as is. 
   
   That said, I had a hard time to fully understand the structure of 
`flink-core`. It would be great if you can help me understand this a bit more. 
I am wondering if the users would have the same questions like below.
   
   - I was initially expecting the function interfaces in 
`org.apache.api.java.functions` path, but there are only `KeySelector` and 
`NullByteKeySelector`. The actual function interfaces are in 
`org.apache.flink.common.functions` package. What is the difference between 
those two packages?
   - Also, I was not quite sure about the differences between 
`org.apache.flink.api.common.io` and `org.apache.flink.core.io`. 
   - In the existing package paths in `flink-core`, the implementation and 
interface classes are put in the same package. This PR introduces a new package 
`lib`. Should the other implementations also be put in a `lib` path?
   - In the existing package paths, `util` packages usually contains internal 
helper classes for the (abstract) implementations of the interfaces. This PR 
seems putting the abstract implementations of the `SourceReader` / `Enumerator` 
interfaces in the `util` package, and expecting other Iterator based sources to 
leverage that? Or are those classes also internal?
   - The follow up question to the above one, it seems that the 
`IteratorSourceReader`/ `IteratorSourceEnumerator` classes are just another 
base implementation of the `Source` interface, why the base implementation of 
the source in `flink-connector-base` was not put into 
`org.apache.flink.api.connector.source.lib.util`? 
   
   Sorry for sticking to discussion, my main concern is that If we ever want to 
revisit the API and package structure, right now might be the best timing, 
because we are essentially introducing new `Source` APIs and will probably 
deprecate old `SourceFunction` and `InputFormat` APIs in some coming release. 
This would be a natural migration towards new APIs if we want to do that. 
Keeping the API / package structure as is and changing them in the future could 
be even more disruptive.
   
   Thanks again.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13040: [FLINK-17073] [checkpointing] checkpointing backpressure if there are too many checkpoints to clean

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13040:
URL: https://github.com/apache/flink/pull/13040#issuecomment-667140037


   
   ## CI report:
   
   * 31abcb25fdf72394dbfa91d0ba9e151968375ad7 UNKNOWN
   * db99952f8166cb4d6337dcb3815e72ff2b49604b UNKNOWN
   * 33af84f70ab92a76818f1a4392fecb0c4a8e62f9 UNKNOWN
   * 12d5698e9185b3b97e5323485ee72617c4e535f9 UNKNOWN
   * b7162b6109b03f905f2c52cc22002b198bc9438e UNKNOWN
   * 5fd3f465ce953ac5944f9f8516259245f50d8f9f Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7086)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7087)
 
   * d83bf413dd7661c4ee87a692b80147f2a6750f46 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7157)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19319) Deprecate StreamExecutionEnvironment.setStreamTimeCharacteristic() and TimeCharacteristic

2020-10-01 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19319?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek closed FLINK-19319.

Fix Version/s: 1.12.0
 Release Note: 
This is formatted in Markdown for easy inclusion in the release post.

In Flink 1.12 the default stream time characteristic has been changed to 
`EventTime`, thus you don't need to call this method for enabling event-time 
support anymore. Explicitly using processing-time windows and timers works in 
event-time mode. If you need to disable watermarks, please use 
`ExecutionConfig.setAutoWatermarkInterval(long)`. If you are using 
`IngestionTime`, please manually set an appropriate `WatermarkStrategy`. If you 
are using generic "time window" operations (for example 
`KeyedStream.timeWindow()` that change behaviour based on the time 
characteristic, please use equivalent operations that explicitly specify 
processing time or event time.
   Resolution: Fixed

master: f8cc82b0c7d3ddd35b17c7f6475b8908363c930a

> Deprecate StreamExecutionEnvironment.setStreamTimeCharacteristic() and 
> TimeCharacteristic
> -
>
> Key: FLINK-19319
> URL: https://issues.apache.org/jira/browse/FLINK-19319
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Aljoscha Krettek
>Assignee: Aljoscha Krettek
>Priority: Major
> Fix For: 1.12.0
>
>
> After FLINK-19317 and FLINK-19318 we don't need this setting anymore. Using 
> (explicit) processing-time windows and processing-time timers work fine in a 
> program that has {{EventTime}} set as a time characteristic and once we 
> deprecate {{timeWindow()}} there are not other operations that change 
> behaviour depending on the time characteristic so there's no need to ever 
> change from the new default of event-time. Similarly, the {{IngestionTime}} 
> setting can be achieved in the future by providing an ingestion-time 
> {{WatermarkStrategy}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13529: [FLINK-19473] Implement multi inputs sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13529:
URL: https://github.com/apache/flink/pull/13529#issuecomment-702193043


   
   ## CI report:
   
   * 6270bf09d731ac358ca621e4150c817153882eb9 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7155)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19317) Make EventTime the default StreamTimeCharacteristic

2020-10-01 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19317?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek closed FLINK-19317.

Fix Version/s: 1.12.0
   Resolution: Fixed

master: 1175364cc9f5bc6018d4c3f231d24cb8ff5e6341

> Make EventTime the default StreamTimeCharacteristic
> ---
>
> Key: FLINK-19317
> URL: https://issues.apache.org/jira/browse/FLINK-19317
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Aljoscha Krettek
>Assignee: Aljoscha Krettek
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> Excerpt from FLIP-134:
> {quote}
> As described above, event time is the only sensible time characteristic for 
> batch. We therefore propose to change the default value of the 
> StreamTimeCharacteristic from ProcessingTime to EventTime. This means the 
> DataStream API programs that were using event time before now just work 
> without manually changing this setting. Processing-time programs will also 
> still work, because using processing-time timers is not dependent on the 
> StreamTimeCharacteristic. DataStream programs that don't set a 
> TimestampAssigner or WatermarkStrategy will also still work if they don't use 
> operations that don't rely on (event-time) timestamps. This is true for both 
> BATCH and STREAMING execution mode.
> The only real user-visible change of this is that programs that used the 
> KeyedStream.timeWindow()/DataStream.timeWindow() operation, which is 
> dependent on the StreamTimeCharacteristic will now use event time by default. 
> We don't think this operation is useful because the behaviour can be 
> surprising. We recommend users always use an explicit processing-time window 
> or event-time window.
> {quote}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] aljoscha commented on pull request #13509: [FLINK-19317] Make EventTime the default TimeCharacteristic

2020-10-01 Thread GitBox


aljoscha commented on pull request #13509:
URL: https://github.com/apache/flink/pull/13509#issuecomment-702205959


   Thanks for the reviews! I merged.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] aljoscha closed pull request #13509: [FLINK-19317] Make EventTime the default TimeCharacteristic

2020-10-01 Thread GitBox


aljoscha closed pull request #13509:
URL: https://github.com/apache/flink/pull/13509


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13040: [FLINK-17073] [checkpointing] checkpointing backpressure if there are too many checkpoints to clean

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13040:
URL: https://github.com/apache/flink/pull/13040#issuecomment-667140037


   
   ## CI report:
   
   * 31abcb25fdf72394dbfa91d0ba9e151968375ad7 UNKNOWN
   * db99952f8166cb4d6337dcb3815e72ff2b49604b UNKNOWN
   * 33af84f70ab92a76818f1a4392fecb0c4a8e62f9 UNKNOWN
   * 12d5698e9185b3b97e5323485ee72617c4e535f9 UNKNOWN
   * b7162b6109b03f905f2c52cc22002b198bc9438e UNKNOWN
   * 5fd3f465ce953ac5944f9f8516259245f50d8f9f Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7086)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7087)
 
   * d83bf413dd7661c4ee87a692b80147f2a6750f46 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-6215) Make the StatefulSequenceSource scalable.

2020-10-01 Thread Kostas Kloudas (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-6215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205583#comment-17205583
 ] 

Kostas Kloudas commented on FLINK-6215:
---

Yes [~sewen], I will close the issue. The PRs are already closed.

> Make the StatefulSequenceSource scalable.
> -
>
> Key: FLINK-6215
> URL: https://issues.apache.org/jira/browse/FLINK-6215
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.3.0
>Reporter: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
>
> Currently the {{StatefulSequenceSource}} instantiates all the elements to 
> emit first and keeps them in memory. This is not scalable as for large 
> sequences of elements this can lead to out of memory exceptions.
> To solve this, we can pre-partition the sequence of elements based on the 
> {{maxParallelism}} parameter, and just keep state (to checkpoint) per such 
> partition.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-6215) Make the StatefulSequenceSource scalable.

2020-10-01 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-6215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-6215.
-
Resolution: Won't Fix

> Make the StatefulSequenceSource scalable.
> -
>
> Key: FLINK-6215
> URL: https://issues.apache.org/jira/browse/FLINK-6215
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.3.0
>Reporter: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
>
> Currently the {{StatefulSequenceSource}} instantiates all the elements to 
> emit first and keeps them in memory. This is not scalable as for large 
> sequences of elements this can lead to out of memory exceptions.
> To solve this, we can pre-partition the sequence of elements based on the 
> {{maxParallelism}} parameter, and just keep state (to checkpoint) per such 
> partition.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] echauchot commented on pull request #13040: [FLINK-17073] [checkpointing] checkpointing backpressure if there are too many checkpoints to clean

2020-10-01 Thread GitBox


echauchot commented on pull request #13040:
URL: https://github.com/apache/flink/pull/13040#issuecomment-702194047


   @tillrohrmann I rebased on master and squashed the PR to only 2 commits:
   - production code + UTest of the CheckpointsDecider
   - integration test
   Can we merge the PR ?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13529: [FLINK-19473] Implement multi inputs sorting DataInput

2020-10-01 Thread GitBox


flinkbot commented on pull request #13529:
URL: https://github.com/apache/flink/pull/13529#issuecomment-702193043


   
   ## CI report:
   
   * 6270bf09d731ac358ca621e4150c817153882eb9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13521:
URL: https://github.com/apache/flink/pull/13521#issuecomment-701282459


   
   ## CI report:
   
   * ad55b5a30711435f6caf0d52ba1cdd702af96fda Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7151)
 
   * c5164bceb05c9e11c652fdb2a0cefe289af0ec1b UNKNOWN
   * 39e66d00bae3c21b41c12881cd2680f74ffbf2e1 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7154)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12056: [FLINK-17502] [flink-connector-rabbitmq] RMQSource refactor

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #12056:
URL: https://github.com/apache/flink/pull/12056#issuecomment-626167900


   
   ## CI report:
   
   * c42ef1f62396948ece865a3b8628bea229d5ffb9 UNKNOWN
   * ad971fd47ab3ac9e813afae9713c75986646 UNKNOWN
   * 4d2d57b8c2f6810c03d284fdd5807b81a2b554d7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7120)
 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7150)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7142)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-6215) Make the StatefulSequenceSource scalable.

2020-10-01 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-6215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205572#comment-17205572
 ] 

Stephan Ewen commented on FLINK-6215:
-

This should issue should be subsumed by this one: FLINK-19457

[~kkl0u] Should we close this and the related PRs?

> Make the StatefulSequenceSource scalable.
> -
>
> Key: FLINK-6215
> URL: https://issues.apache.org/jira/browse/FLINK-6215
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.3.0
>Reporter: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
>
> Currently the {{StatefulSequenceSource}} instantiates all the elements to 
> emit first and keeps them in memory. This is not scalable as for large 
> sequences of elements this can lead to out of memory exceptions.
> To solve this, we can pre-partition the sequence of elements based on the 
> {{maxParallelism}} parameter, and just keep state (to checkpoint) per such 
> partition.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13521:
URL: https://github.com/apache/flink/pull/13521#issuecomment-701282459


   
   ## CI report:
   
   * ad55b5a30711435f6caf0d52ba1cdd702af96fda Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7151)
 
   * c5164bceb05c9e11c652fdb2a0cefe289af0ec1b UNKNOWN
   * 39e66d00bae3c21b41c12881cd2680f74ffbf2e1 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #13529: [FLINK-19473] Implement multi inputs sorting DataInput

2020-10-01 Thread GitBox


flinkbot commented on pull request #13529:
URL: https://github.com/apache/flink/pull/13529#issuecomment-702178922


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 6270bf09d731ac358ca621e4150c817153882eb9 (Thu Oct 01 
14:35:15 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-19473) Implement multi inputs sorting DataInput

2020-10-01 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-19473:
---
Labels: pull-request-available  (was: )

> Implement multi inputs sorting DataInput
> 
>
> Key: FLINK-19473
> URL: https://issues.apache.org/jira/browse/FLINK-19473
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Task
>Reporter: Dawid Wysakowicz
>Assignee: Dawid Wysakowicz
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19447) HBaseConnectorITCase.HBaseTestingClusterAutoStarter failed with "Master not initialized after 200000ms"

2020-10-01 Thread Miklos Gergely (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205566#comment-17205566
 ] 

Miklos Gergely commented on FLINK-19447:


Thank you [~dian.fu]. Created https://issues.apache.org/jira/browse/HBASE-25140 
to track the issue of the test mini cluster compatibility.

> HBaseConnectorITCase.HBaseTestingClusterAutoStarter failed with "Master not 
> initialized after 20ms"
> ---
>
> Key: FLINK-19447
> URL: https://issues.apache.org/jira/browse/FLINK-19447
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase, Tests
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Assignee: Miklos Gergely
>Priority: Blocker
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
> Attachments: screenshot-1.png, screenshot-2.png, screenshot-3.png
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7042=logs=961f8f81-6b52-53df-09f6-7291a2e4af6a=60581941-0138-53c0-39fe-86d62be5f407
> {code}
> 2020-09-28T21:52:21.2146147Z 
> org.apache.flink.connector.hbase2.HBaseConnectorITCase  Time elapsed: 208.382 
> sec  <<< ERROR!
> 2020-09-28T21:52:21.2146638Z java.io.IOException: Shutting down
> 2020-09-28T21:52:21.2147004Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:266)
> 2020-09-28T21:52:21.2147637Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.(MiniHBaseCluster.java:116)
> 2020-09-28T21:52:21.2148120Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniHBaseCluster(HBaseTestingUtility.java:1142)
> 2020-09-28T21:52:21.2148831Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1107)
> 2020-09-28T21:52:21.2149347Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1061)
> 2020-09-28T21:52:21.2149896Z  at 
> org.apache.flink.connector.hbase2.util.HBaseTestingClusterAutoStarter.setUp(HBaseTestingClusterAutoStarter.java:122)
> 2020-09-28T21:52:21.2150721Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-09-28T21:52:21.2151136Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-09-28T21:52:21.2151609Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-09-28T21:52:21.2152039Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-09-28T21:52:21.2152462Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-09-28T21:52:21.2152941Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-09-28T21:52:21.2153489Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-09-28T21:52:21.2153962Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
> 2020-09-28T21:52:21.2154406Z  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2020-09-28T21:52:21.2154828Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-09-28T21:52:21.2155381Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:367)
> 2020-09-28T21:52:21.2155864Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:274)
> 2020-09-28T21:52:21.2156378Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> 2020-09-28T21:52:21.2156865Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:161)
> 2020-09-28T21:52:21.2157458Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:290)
> 2020-09-28T21:52:21.2157993Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:242)
> 2020-09-28T21:52:21.2158470Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:121)
> 2020-09-28T21:52:21.2158890Z Caused by: java.lang.RuntimeException: Master 
> not initialized after 20ms
> 2020-09-28T21:52:21.2159350Z  at 
> org.apache.hadoop.hbase.util.JVMClusterUtil.waitForEvent(JVMClusterUtil.java:229)
> 2020-09-28T21:52:21.2159823Z  at 
> org.apache.hadoop.hbase.util.JVMClusterUtil.startup(JVMClusterUtil.java:197)
> 2020-09-28T21:52:21.2160270Z  at 
> org.apache.hadoop.hbase.LocalHBaseCluster.startup(LocalHBaseCluster.java:413)
> 2020-09-28T21:52:21.2160800Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:259)
> 2020-09-28T21:52:21.2161096Z  ... 22 more
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] dawidwys opened a new pull request #13529: [FLINK-19473] Implement multi inputs sorting DataInput

2020-10-01 Thread GitBox


dawidwys opened a new pull request #13529:
URL: https://github.com/apache/flink/pull/13529


   ## What is the purpose of the change
   
   The PR is still rough on the edges. It requires updates to the commit 
messages. I am opening the PR to get early feedback.
   Moreover it is based on #13521
   
   
   ## Brief change log
   See commits subjects for the list of changes.
   
   
   ## Verifying this change
   
   Added tests in:
   * MultiInputSortingDataInputsTest
   * SortedDataInputITCase
   * SortingBoundedInputITCase.java 
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (**yes** / no 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes / **no** / 
don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-19487) Checkpoint start delay is always zero for single channel tasks

2020-10-01 Thread Piotr Nowojski (Jira)
Piotr Nowojski created FLINK-19487:
--

 Summary: Checkpoint start delay is always zero for single channel 
tasks
 Key: FLINK-19487
 URL: https://issues.apache.org/jira/browse/FLINK-19487
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Metrics
Affects Versions: 1.11.2
Reporter: Piotr Nowojski
Assignee: Piotr Nowojski
 Fix For: 1.12.0


In {{CheckpointBarrierAligner}} we are marking start of the checkpoint only if 
the number of channels is more than 1. As a result of this, 
{{checkpointStartDelay}} is never updated.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] StephanEwen commented on pull request #13501: Single task result partition type

2020-10-01 Thread GitBox


StephanEwen commented on pull request #13501:
URL: https://github.com/apache/flink/pull/13501#issuecomment-702157253


   I took a look only at the partial record handling. I like the idea of not 
transporting the partial record offsets at all across the networks, which means 
we don't have to make the deserializers aware of that.
   
   Regarding the implementation, I have the same comment as @pnowojski . Given 
that the partial lengths fields are not communicated across the network, I feel 
that this logic does not belong into buffers or 
`BufferBuilder`/`BufferConsumer`. Those should handle purely bytes without any 
knowledge of records.
   
   I would suggest to change the `buffers` queue from holding `BufferConsumer` 
to holding `BufferConsumerWithOffsets` or so (maybe come up with a shorter 
name). That way we transport the same information, but not inject it into the 
buffers.
   
   Some more things to watch out for in further refactoring:
   
 - The *per record methods* of *ResultPartition* and its subclasses are 
highly performance sensitive. They should not become virtual methods. We 
currently can guarantee that because there is only one class implementing these 
per-record methods (BufferWritingResultPartition).
   When we start having different implementations, we ideally have only one 
of them loaded in the JVM at any point, so that the de-virtualization works 
properly. That means design the classes/subclasses such that a type of job 
(batch, streaming, streaming with partial failover) would only ever rely on one 
of these partition types in a job.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Comment Edited] (FLINK-19332) Special characters issue using Kinesis Data Analytics for Apache Flink

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205532#comment-17205532
 ] 

Dian Fu edited comment on FLINK-19332 at 10/1/20, 1:50 PM:
---

Hi [~zekuny] Personally I guess this should not be a *Blocker* issue. Could you 
explain why you think this is a *Blocker* issue?


was (Author: dian.fu):
Hi [~zekuny] Personally I guess this should not be a **Blocker** issue. Could 
you explain why you think this is a **Blocker** issue?

> Special characters issue using Kinesis Data Analytics for Apache Flink
> --
>
> Key: FLINK-19332
> URL: https://issues.apache.org/jira/browse/FLINK-19332
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.8.2
>Reporter: Zekun Yu
>Priority: Blocker
> Fix For: 1.8.2
>
>
> Hi there,
>  
> I am encountering one special character issue while using Kinesis Data 
> Analytics for Apache Flink (KDA).
>  
> Our KDA is built for processing data and outputting to a Kinesis stream. We 
> have a lambda function that subscribes to the Kinesis stream and reads 
> records from the Kinesis stream.
> The library in the KDA I am using is 
> "org.apache.flink.streaming.connectors.kinesis".
>  
> Our KDA is only outputting one single record to the Kinesis sink using 
> "collector.collect()" for a single key (details will be found below)
> Most times, the record received by the Lambda looks perfectly good.
> However, occasionally, when two records are sent to the kinesis sink using 
> "collector.collect()" at the same time, *we noticed that those two records 
> are combined somehow and there are some special characters in the record 
> received by the Lambda function*.
>  
>  
>  
>  
>  
>  
> Below are some technical details:
>  
> The KDA is not using any "TimeWindow()" but uses "keyBy()" by some keys.
> ).returns(MatchedDataForAlarm.class)
>     .keyBy(MatchedDataForAlarm::getStateKey)
>     .connect(ruleBroadcastStream)
>     .process(new MetricProcess())
>     .addSink(kinesis);
>  
>  
> The "MetricProcess()" extends "KeyedBroadcastProcessFunction" which overrides 
> the "processElement" function. It uses collector.collect() for outputs.
> @Override
> public void processElement(MatchedDataForAlarm input, ReadOnlyContext 
> ctx,Collector collector) throws Exception {
>  
>  
> We have our own AEMMatchedDataForAlarmSchemaSerialization which implements 
> KinesisSerializationSchema. The serialization simply 
> converts a "MatchedDataForAlarm" object to String using Gson and then 
> converts to ByteBuffer.
>  
> @Override
> public ByteBuffer serialize(MatchedDataForAlarm matchedDataForAlarm) {
>     Gson gson = new Gson();
> String result = gson.toJson(matchedDataForAlarm);
> _log_.info("Alarm record sent to Kinesis stream: {}", result);
>     return ByteBuffer._wrap_(result.getBytes());
> }
>  
>  
>  
> *Here's the record shown in the Lambda logs when two records are combined 
> somewhere somehow (most cases those two are received as two separate 
> records):*
>  
> 
> 0??
> { "inAlarmState": false }
> ??
> { "inAlarmState": false}
> eEo?N9x
>  
>  
>  
>  
> I am not sure if it's a serialization issue or some default behaviors in the 
> Kinesis sink library? It might be just some common mistakes that I made which 
> I am not aware of.
> Could anyone help with this problem? I really appreciate it.
>  
>  
>  
> Thanks,
> Zekun
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19332) Special characters issue using Kinesis Data Analytics for Apache Flink

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205532#comment-17205532
 ] 

Dian Fu commented on FLINK-19332:
-

Hi [~zekuny] Personally I guess this should not be a **Blocker** issue. Could 
you explain why you think this is a **Blocker** issue?

> Special characters issue using Kinesis Data Analytics for Apache Flink
> --
>
> Key: FLINK-19332
> URL: https://issues.apache.org/jira/browse/FLINK-19332
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.8.2
>Reporter: Zekun Yu
>Priority: Blocker
> Fix For: 1.8.2
>
>
> Hi there,
>  
> I am encountering one special character issue while using Kinesis Data 
> Analytics for Apache Flink (KDA).
>  
> Our KDA is built for processing data and outputting to a Kinesis stream. We 
> have a lambda function that subscribes to the Kinesis stream and reads 
> records from the Kinesis stream.
> The library in the KDA I am using is 
> "org.apache.flink.streaming.connectors.kinesis".
>  
> Our KDA is only outputting one single record to the Kinesis sink using 
> "collector.collect()" for a single key (details will be found below)
> Most times, the record received by the Lambda looks perfectly good.
> However, occasionally, when two records are sent to the kinesis sink using 
> "collector.collect()" at the same time, *we noticed that those two records 
> are combined somehow and there are some special characters in the record 
> received by the Lambda function*.
>  
>  
>  
>  
>  
>  
> Below are some technical details:
>  
> The KDA is not using any "TimeWindow()" but uses "keyBy()" by some keys.
> ).returns(MatchedDataForAlarm.class)
>     .keyBy(MatchedDataForAlarm::getStateKey)
>     .connect(ruleBroadcastStream)
>     .process(new MetricProcess())
>     .addSink(kinesis);
>  
>  
> The "MetricProcess()" extends "KeyedBroadcastProcessFunction" which overrides 
> the "processElement" function. It uses collector.collect() for outputs.
> @Override
> public void processElement(MatchedDataForAlarm input, ReadOnlyContext 
> ctx,Collector collector) throws Exception {
>  
>  
> We have our own AEMMatchedDataForAlarmSchemaSerialization which implements 
> KinesisSerializationSchema. The serialization simply 
> converts a "MatchedDataForAlarm" object to String using Gson and then 
> converts to ByteBuffer.
>  
> @Override
> public ByteBuffer serialize(MatchedDataForAlarm matchedDataForAlarm) {
>     Gson gson = new Gson();
> String result = gson.toJson(matchedDataForAlarm);
> _log_.info("Alarm record sent to Kinesis stream: {}", result);
>     return ByteBuffer._wrap_(result.getBytes());
> }
>  
>  
>  
> *Here's the record shown in the Lambda logs when two records are combined 
> somewhere somehow (most cases those two are received as two separate 
> records):*
>  
> 
> 0??
> { "inAlarmState": false }
> ??
> { "inAlarmState": false}
> eEo?N9x
>  
>  
>  
>  
> I am not sure if it's a serialization issue or some default behaviors in the 
> Kinesis sink library? It might be just some common mistakes that I made which 
> I am not aware of.
> Could anyone help with this problem? I really appreciate it.
>  
>  
>  
> Thanks,
> Zekun
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19486) expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle

2020-10-01 Thread appleyuchi (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

appleyuchi updated FLINK-19486:
---
Description: 
*I want to do an experiment of"incremental checkpoint"*

my code is:

[https://paste.ubuntu.com/p/DpTyQKq6Vk/]

 

pom.xml is:


 http://maven.apache.org/POM/4.0.0;
 xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
 xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
[http://maven.apache.org/xsd/maven-4.0.0.xsd];>
 4.0.0

example
 datastream_api
 1.0-SNAPSHOT
 
 
 
 org.apache.maven.plugins
 maven-compiler-plugin
 3.1
 
 1.8
 1.8
 
 


 org.scala-tools
 maven-scala-plugin
 2.15.2
 
 
 
 compile
 testCompile
 
 
 
 

 


 




 
 org.apache.flink
 flink-streaming-scala_2.11
 1.11.1
 
 


 
 
 
 
 


 org.apache.flink
 flink-clients_2.11
 1.11.1
 

 


 org.apache.flink
 flink-statebackend-rocksdb_2.11
 1.11.2
 
 


 org.apache.hadoop
 hadoop-client
 3.3.0
 


 org.apache.flink
 flink-core
 1.11.1
 


 
 
 
 
 


 
 org.apache.flink
 flink-cep_2.11
 1.11.1
 


 org.apache.flink
 flink-cep-scala_2.11
 1.11.1
 


 org.apache.flink
 flink-scala_2.11
 1.11.1
 

 


 org.projectlombok
 lombok
 1.18.4
 
 


 

 

the error I got is:

[https://paste.ubuntu.com/p/49HRYXFzR2/]

 

*some of the above error is:*

*Caused by: java.lang.IllegalStateException: Unexpected state handle type, 
expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: 
class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle*

 

 

The steps are:

1.mvn clean scala:compile compile package

2.nc -lk 

3.flink run -c wordcount_increstate  datastream_api-1.0-SNAPSHOT.jar
 Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b

4.input the following conents in nc -lk 

before
 error
 error
 error
 error

5.

flink run -s 
hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c 
StateWordCount datastream_api-1.0-SNAPSHOT.jar

Then the above error happens.

 

Please help,Thanks~!

 

  was:
I want to do an experiment of"incremental checkpoint"

my code is:

[https://paste.ubuntu.com/p/DpTyQKq6Vk/]

 

pom.xml is:


http://maven.apache.org/POM/4.0.0;
 xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
 xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
 4.0.0

example
 datastream_api
 1.0-SNAPSHOT
 
 
 
 org.apache.maven.plugins
 maven-compiler-plugin
 3.1
 
 1.8
 1.8
 
 


 
 org.scala-tools
 maven-scala-plugin
 2.15.2
 
 
 
 compile
 testCompile
 
 
 
 

 


 


 


 
 org.apache.flink
 flink-streaming-scala_2.11
 1.11.1
 
 


 
 
 
 
 


 org.apache.flink
 flink-clients_2.11
 1.11.1
 

 


 org.apache.flink
 flink-statebackend-rocksdb_2.11
 1.11.2
 
 


 
 org.apache.hadoop
 hadoop-client
 3.3.0
 


 
 org.apache.flink
 flink-core
 1.11.1
 


 
 
 
 
 
 


 
 
 org.apache.flink
 flink-cep_2.11
 1.11.1
 


 
 org.apache.flink
 flink-cep-scala_2.11
 1.11.1
 


 
 org.apache.flink
 flink-scala_2.11
 1.11.1
 

 


 org.projectlombok
 lombok
 1.18.4
 
 


 


 

the error I got is:

[https://paste.ubuntu.com/p/49HRYXFzR2/]

 

*some of the above error is:*

*Caused by: java.lang.IllegalStateException: Unexpected state handle type, 
expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: 
class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle*

 

 

The steps are:

1.mvn clean scala:compile compile package

2.nc -lk 

3.flink run -c wordcount_increstate  datastream_api-1.0-SNAPSHOT.jar
Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b

4.input the following conents in nc -lk 

before
error
error
error
error

5.

flink run -s 
hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c 
StateWordCount datastream_api-1.0-SNAPSHOT.jar

Then the above error happens.

 

Please help,Thanks~!

 


> expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but 
> found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle
> --
>
> Key: FLINK-19486
> URL: https://issues.apache.org/jira/browse/FLINK-19486
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.11.1
>Reporter: appleyuchi
>Priority: Blocker
> Fix For: 1.11.1
>
>
> *I want to do an experiment of"incremental checkpoint"*
> my code is:
> [https://paste.ubuntu.com/p/DpTyQKq6Vk/]
>  
> pom.xml is:
> 
>  http://maven.apache.org/POM/4.0.0;
>  xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
>  xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
> [http://maven.apache.org/xsd/maven-4.0.0.xsd];>
>  4.0.0
> example
>  datastream_api
>  1.0-SNAPSHOT
>  
>  
>  
>  org.apache.maven.plugins
>  

[jira] [Updated] (FLINK-19486) expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle

2020-10-01 Thread appleyuchi (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

appleyuchi updated FLINK-19486:
---
Issue Type: Bug  (was: Test)

> expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but 
> found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle
> --
>
> Key: FLINK-19486
> URL: https://issues.apache.org/jira/browse/FLINK-19486
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.11.1
>Reporter: appleyuchi
>Priority: Blocker
> Fix For: 1.11.1
>
>
> I want to do an experiment of"incremental checkpoint"
> my code is:
> [https://paste.ubuntu.com/p/DpTyQKq6Vk/]
>  
> pom.xml is:
> 
> http://maven.apache.org/POM/4.0.0;
>  xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
>  xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
> http://maven.apache.org/xsd/maven-4.0.0.xsd;>
>  4.0.0
> example
>  datastream_api
>  1.0-SNAPSHOT
>  
>  
>  
>  org.apache.maven.plugins
>  maven-compiler-plugin
>  3.1
>  
>  1.8
>  1.8
>  
>  
>  
>  org.scala-tools
>  maven-scala-plugin
>  2.15.2
>  
>  
>  
>  compile
>  testCompile
>  
>  
>  
>  
>  
> 
>  
>  
> 
>  
>  org.apache.flink
>  flink-streaming-scala_2.11
>  1.11.1
>  
>  
> 
>  
>  
>  
>  
>  
> 
>  org.apache.flink
>  flink-clients_2.11
>  1.11.1
>  
>  
> 
>  org.apache.flink
>  flink-statebackend-rocksdb_2.11
>  1.11.2
>  
>  
>  
>  org.apache.hadoop
>  hadoop-client
>  3.3.0
>  
>  
>  org.apache.flink
>  flink-core
>  1.11.1
>  
>  
>  
>  
>  
>  
>  
>  
>  
>  org.apache.flink
>  flink-cep_2.11
>  1.11.1
>  
>  
>  org.apache.flink
>  flink-cep-scala_2.11
>  1.11.1
>  
>  
>  org.apache.flink
>  flink-scala_2.11
>  1.11.1
>  
>  
> 
>  org.projectlombok
>  lombok
>  1.18.4
>  
>  
>  
> 
>  
> the error I got is:
> [https://paste.ubuntu.com/p/49HRYXFzR2/]
>  
> *some of the above error is:*
> *Caused by: java.lang.IllegalStateException: Unexpected state handle type, 
> expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but 
> found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle*
>  
>  
> The steps are:
> 1.mvn clean scala:compile compile package
> 2.nc -lk 
> 3.flink run -c wordcount_increstate  datastream_api-1.0-SNAPSHOT.jar
> Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b
> 4.input the following conents in nc -lk 
> before
> error
> error
> error
> error
> 5.
> flink run -s 
> hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c 
> StateWordCount datastream_api-1.0-SNAPSHOT.jar
> Then the above error happens.
>  
> Please help,Thanks~!
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-19486) expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle

2020-10-01 Thread appleyuchi (Jira)
appleyuchi created FLINK-19486:
--

 Summary: expected: class 
org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: class 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle
 Key: FLINK-19486
 URL: https://issues.apache.org/jira/browse/FLINK-19486
 Project: Flink
  Issue Type: Test
Affects Versions: 1.11.1
Reporter: appleyuchi
 Fix For: 1.11.1


I want to do an experiment of"incremental checkpoint"

my code is:

[https://paste.ubuntu.com/p/DpTyQKq6Vk/]

 

pom.xml is:


http://maven.apache.org/POM/4.0.0;
 xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance;
 xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd;>
 4.0.0

example
 datastream_api
 1.0-SNAPSHOT
 
 
 
 org.apache.maven.plugins
 maven-compiler-plugin
 3.1
 
 1.8
 1.8
 
 


 
 org.scala-tools
 maven-scala-plugin
 2.15.2
 
 
 
 compile
 testCompile
 
 
 
 

 


 


 


 
 org.apache.flink
 flink-streaming-scala_2.11
 1.11.1
 
 


 
 
 
 
 


 org.apache.flink
 flink-clients_2.11
 1.11.1
 

 


 org.apache.flink
 flink-statebackend-rocksdb_2.11
 1.11.2
 
 


 
 org.apache.hadoop
 hadoop-client
 3.3.0
 


 
 org.apache.flink
 flink-core
 1.11.1
 


 
 
 
 
 
 


 
 
 org.apache.flink
 flink-cep_2.11
 1.11.1
 


 
 org.apache.flink
 flink-cep-scala_2.11
 1.11.1
 


 
 org.apache.flink
 flink-scala_2.11
 1.11.1
 

 


 org.projectlombok
 lombok
 1.18.4
 
 


 


 

the error I got is:

[https://paste.ubuntu.com/p/49HRYXFzR2/]

 

*some of the above error is:*

*Caused by: java.lang.IllegalStateException: Unexpected state handle type, 
expected: class org.apache.flink.runtime.state.KeyGroupsStateHandle, but found: 
class org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle*

 

 

The steps are:

1.mvn clean scala:compile compile package

2.nc -lk 

3.flink run -c wordcount_increstate  datastream_api-1.0-SNAPSHOT.jar
Job has been submitted with JobID df6d62a43620f258155b8538ef0ddf1b

4.input the following conents in nc -lk 

before
error
error
error
error

5.

flink run -s 
hdfs://Desktop:9000/tmp/flinkck/df6d62a43620f258155b8538ef0ddf1b/chk-22 -c 
StateWordCount datastream_api-1.0-SNAPSHOT.jar

Then the above error happens.

 

Please help,Thanks~!

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19484) Kubernetes pyflink application test failed with "error executing jsonpath "{range .items[*]}{.metadata.name}{\"\\n\"}{end}": Error executing template: not in rang"

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19484?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205514#comment-17205514
 ] 

Dian Fu commented on FLINK-19484:
-

Another instance: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7148=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529

> Kubernetes pyflink application test failed with "error executing jsonpath 
> "{range .items[*]}{.metadata.name}{\"\\n\"}{end}": Error executing template: 
> not in rang"
> ---
>
> Key: FLINK-19484
> URL: https://issues.apache.org/jira/browse/FLINK-19484
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Tests
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7139=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> 2020-09-30T21:14:41.0570715Z 2020-09-30 21:14:41,056 INFO  
> org.apache.flink.kubernetes.KubernetesClusterDescriptor  [] - Create 
> flink application cluster flink-native-k8s-pyflink-application-1 
> successfully, JobManager Web Interface: http://10.1.0.4:30141
> 2020-09-30T21:15:11.2323195Z error: timed out waiting for the condition on 
> deployments/flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.2384760Z Stopping job timeout watchdog (with pid=111683)
> 2020-09-30T21:15:11.2386729Z Debugging failed Kubernetes test:
> 2020-09-30T21:15:11.2387191Z Currently existing Kubernetes resources
> 2020-09-30T21:15:11.3502610Z NAME 
>  TYPECLUSTER-IP EXTERNAL-IP   PORT(S) AGE
> 2020-09-30T21:15:11.3506194Z service/flink-native-k8s-pyflink-application-1   
>  ClusterIP   None   6123/TCP,6124/TCP   31s
> 2020-09-30T21:15:11.3507403Z 
> service/flink-native-k8s-pyflink-application-1-rest   NodePort
> 10.104.215.1   8081:30141/TCP  31s
> 2020-09-30T21:15:11.3529743Z service/kubernetes   
>  ClusterIP   10.96.0.1  443/TCP 18m
> 2020-09-30T21:15:11.3530391Z 
> 2020-09-30T21:15:11.3531349Z NAME 
> READY   UP-TO-DATE   AVAILABLE   AGE
> 2020-09-30T21:15:11.3532200Z 
> deployment.apps/flink-native-k8s-pyflink-application-1   0/1 0
> 0   31s
> 2020-09-30T21:15:11.4887105Z Name:  
> flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4887491Z Namespace: default
> 2020-09-30T21:15:11.4888028Z Labels:
> app=flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4888534Ztype=flink-native-kubernetes
> 2020-09-30T21:15:11.443Z Annotations:   
> 2020-09-30T21:15:11.4890558Z Selector:  
> app=flink-native-k8s-pyflink-application-1,component=jobmanager,type=flink-native-kubernetes
> 2020-09-30T21:15:11.4900945Z Type:  ClusterIP
> 2020-09-30T21:15:11.4901267Z IP:None
> 2020-09-30T21:15:11.4903601Z Port:  jobmanager-rpc  6123/TCP
> 2020-09-30T21:15:11.4903911Z TargetPort:6123/TCP
> 2020-09-30T21:15:11.4904155Z Endpoints: 
> 2020-09-30T21:15:11.4907175Z Port:  blobserver  6124/TCP
> 2020-09-30T21:15:11.4907586Z TargetPort:6124/TCP
> 2020-09-30T21:15:11.4907842Z Endpoints: 
> 2020-09-30T21:15:11.4908063Z Session Affinity:  None
> 2020-09-30T21:15:11.4908298Z Events:
> 2020-09-30T21:15:11.4970434Z 
> 2020-09-30T21:15:11.4970653Z 
> 2020-09-30T21:15:11.4971488Z Name: 
> flink-native-k8s-pyflink-application-1-rest
> 2020-09-30T21:15:11.4971843Z Namespace:default
> 2020-09-30T21:15:11.4972563Z Labels:   
> app=flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4973297Z   
> type=flink-native-kubernetes
> 2020-09-30T21:15:11.4973749Z Annotations:  
> 2020-09-30T21:15:11.4974568Z Selector: 
> app=flink-native-k8s-pyflink-application-1,component=jobmanager,type=flink-native-kubernetes
> 2020-09-30T21:15:11.4974969Z Type: NodePort
> 2020-09-30T21:15:11.4976265Z IP:   10.104.215.1
> 2020-09-30T21:15:11.4976620Z Port: rest  8081/TCP
> 2020-09-30T21:15:11.4976991Z TargetPort:   8081/TCP
> 2020-09-30T21:15:11.4977280Z NodePort: rest  30141/TCP
> 2020-09-30T21:15:11.4977860Z Endpoints:
> 2020-09-30T21:15:11.4978117Z Session Affinity: None
> 

[jira] [Updated] (FLINK-19484) Kubernetes pyflink application test failed with "error executing jsonpath "{range .items[*]}{.metadata.name}{\"\\n\"}{end}": Error executing template: not in rang"

2020-10-01 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-19484:

Priority: Critical  (was: Major)

> Kubernetes pyflink application test failed with "error executing jsonpath 
> "{range .items[*]}{.metadata.name}{\"\\n\"}{end}": Error executing template: 
> not in rang"
> ---
>
> Key: FLINK-19484
> URL: https://issues.apache.org/jira/browse/FLINK-19484
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Tests
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7139=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=ff888d9b-cd34-53cc-d90f-3e446d355529
> {code}
> 2020-09-30T21:14:41.0570715Z 2020-09-30 21:14:41,056 INFO  
> org.apache.flink.kubernetes.KubernetesClusterDescriptor  [] - Create 
> flink application cluster flink-native-k8s-pyflink-application-1 
> successfully, JobManager Web Interface: http://10.1.0.4:30141
> 2020-09-30T21:15:11.2323195Z error: timed out waiting for the condition on 
> deployments/flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.2384760Z Stopping job timeout watchdog (with pid=111683)
> 2020-09-30T21:15:11.2386729Z Debugging failed Kubernetes test:
> 2020-09-30T21:15:11.2387191Z Currently existing Kubernetes resources
> 2020-09-30T21:15:11.3502610Z NAME 
>  TYPECLUSTER-IP EXTERNAL-IP   PORT(S) AGE
> 2020-09-30T21:15:11.3506194Z service/flink-native-k8s-pyflink-application-1   
>  ClusterIP   None   6123/TCP,6124/TCP   31s
> 2020-09-30T21:15:11.3507403Z 
> service/flink-native-k8s-pyflink-application-1-rest   NodePort
> 10.104.215.1   8081:30141/TCP  31s
> 2020-09-30T21:15:11.3529743Z service/kubernetes   
>  ClusterIP   10.96.0.1  443/TCP 18m
> 2020-09-30T21:15:11.3530391Z 
> 2020-09-30T21:15:11.3531349Z NAME 
> READY   UP-TO-DATE   AVAILABLE   AGE
> 2020-09-30T21:15:11.3532200Z 
> deployment.apps/flink-native-k8s-pyflink-application-1   0/1 0
> 0   31s
> 2020-09-30T21:15:11.4887105Z Name:  
> flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4887491Z Namespace: default
> 2020-09-30T21:15:11.4888028Z Labels:
> app=flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4888534Ztype=flink-native-kubernetes
> 2020-09-30T21:15:11.443Z Annotations:   
> 2020-09-30T21:15:11.4890558Z Selector:  
> app=flink-native-k8s-pyflink-application-1,component=jobmanager,type=flink-native-kubernetes
> 2020-09-30T21:15:11.4900945Z Type:  ClusterIP
> 2020-09-30T21:15:11.4901267Z IP:None
> 2020-09-30T21:15:11.4903601Z Port:  jobmanager-rpc  6123/TCP
> 2020-09-30T21:15:11.4903911Z TargetPort:6123/TCP
> 2020-09-30T21:15:11.4904155Z Endpoints: 
> 2020-09-30T21:15:11.4907175Z Port:  blobserver  6124/TCP
> 2020-09-30T21:15:11.4907586Z TargetPort:6124/TCP
> 2020-09-30T21:15:11.4907842Z Endpoints: 
> 2020-09-30T21:15:11.4908063Z Session Affinity:  None
> 2020-09-30T21:15:11.4908298Z Events:
> 2020-09-30T21:15:11.4970434Z 
> 2020-09-30T21:15:11.4970653Z 
> 2020-09-30T21:15:11.4971488Z Name: 
> flink-native-k8s-pyflink-application-1-rest
> 2020-09-30T21:15:11.4971843Z Namespace:default
> 2020-09-30T21:15:11.4972563Z Labels:   
> app=flink-native-k8s-pyflink-application-1
> 2020-09-30T21:15:11.4973297Z   
> type=flink-native-kubernetes
> 2020-09-30T21:15:11.4973749Z Annotations:  
> 2020-09-30T21:15:11.4974568Z Selector: 
> app=flink-native-k8s-pyflink-application-1,component=jobmanager,type=flink-native-kubernetes
> 2020-09-30T21:15:11.4974969Z Type: NodePort
> 2020-09-30T21:15:11.4976265Z IP:   10.104.215.1
> 2020-09-30T21:15:11.4976620Z Port: rest  8081/TCP
> 2020-09-30T21:15:11.4976991Z TargetPort:   8081/TCP
> 2020-09-30T21:15:11.4977280Z NodePort: rest  30141/TCP
> 2020-09-30T21:15:11.4977860Z Endpoints:
> 2020-09-30T21:15:11.4978117Z Session Affinity: None
> 2020-09-30T21:15:11.4978374Z External Traffic Policy:  Cluster
> 2020-09-30T21:15:11.4978719Z Events:   
> 2020-09-30T21:15:11.5048919Z 
> 

[jira] [Closed] (FLINK-19264) Jobs with identical graph shapes cannot be run concurrently

2020-10-01 Thread Zhu Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19264?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhu Zhu closed FLINK-19264.
---
Resolution: Fixed

Fixed via 6e5240f8ac24974826a07b5370908a85bd874ce1

> Jobs with identical graph shapes cannot be run concurrently
> ---
>
> Key: FLINK-19264
> URL: https://issues.apache.org/jira/browse/FLINK-19264
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.12.0
>Reporter: Aljoscha Krettek
>Assignee: Yangze Guo
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> While working on FLINK-19123 I noticed that jobs often fail on 
> {{MiniCluster}} when multiple jobs are running at the same time. 
> I created a reproducer here: 
> https://github.com/aljoscha/flink/tree/flink-19123-fix-test-stream-env-alternative.
>  You can run {{MiniClusterConcurrencyITCase}} to see the problem in action. 
> Sometimes the test will succeed, sometimes it will fail with
> {code}
> java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
>   at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
>   at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
>   at 
> org.apache.flink.test.example.MiniClusterConcurrencyITCase.submitConcurrently(MiniClusterConcurrencyITCase.java:60)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
>   at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
>   at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>   at 
> com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:69)
>   at 
> com.intellij.rt.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:33)
>   at 
> com.intellij.rt.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:220)
>   at com.intellij.rt.junit.JUnitStarter.main(JUnitStarter.java:53)
> Caused by: org.apache.flink.runtime.client.JobExecutionException: Job 
> execution failed.
>   at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
>   at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$2(MiniClusterJobClient.java:107)
>   at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
>   at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229)
>   at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
>   at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:926)
>   at akka.dispatch.OnComplete.internal(Future.scala:264)
>   at 

[GitHub] [flink] zhuzhurk merged pull request #13473: [FLINK-19264][runtime] Add JobID to ExecutionAttemptID

2020-10-01 Thread GitBox


zhuzhurk merged pull request #13473:
URL: https://github.com/apache/flink/pull/13473


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] zhuzhurk commented on pull request #13473: [FLINK-19264][runtime] Add JobID to ExecutionAttemptID

2020-10-01 Thread GitBox


zhuzhurk commented on pull request #13473:
URL: https://github.com/apache/flink/pull/13473#issuecomment-702137444


   Thanks for further addressing the comments. @KarmaGYZ 
   Merging.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Closed] (FLINK-19447) HBaseConnectorITCase.HBaseTestingClusterAutoStarter failed with "Master not initialized after 200000ms"

2020-10-01 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu closed FLINK-19447.
---
Resolution: Fixed

Fixed in master via c05027370b33c178d8b5d597b80a98857c941ba0

> HBaseConnectorITCase.HBaseTestingClusterAutoStarter failed with "Master not 
> initialized after 20ms"
> ---
>
> Key: FLINK-19447
> URL: https://issues.apache.org/jira/browse/FLINK-19447
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase, Tests
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Assignee: Miklos Gergely
>Priority: Blocker
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
> Attachments: screenshot-1.png, screenshot-2.png, screenshot-3.png
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=7042=logs=961f8f81-6b52-53df-09f6-7291a2e4af6a=60581941-0138-53c0-39fe-86d62be5f407
> {code}
> 2020-09-28T21:52:21.2146147Z 
> org.apache.flink.connector.hbase2.HBaseConnectorITCase  Time elapsed: 208.382 
> sec  <<< ERROR!
> 2020-09-28T21:52:21.2146638Z java.io.IOException: Shutting down
> 2020-09-28T21:52:21.2147004Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:266)
> 2020-09-28T21:52:21.2147637Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.(MiniHBaseCluster.java:116)
> 2020-09-28T21:52:21.2148120Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniHBaseCluster(HBaseTestingUtility.java:1142)
> 2020-09-28T21:52:21.2148831Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1107)
> 2020-09-28T21:52:21.2149347Z  at 
> org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:1061)
> 2020-09-28T21:52:21.2149896Z  at 
> org.apache.flink.connector.hbase2.util.HBaseTestingClusterAutoStarter.setUp(HBaseTestingClusterAutoStarter.java:122)
> 2020-09-28T21:52:21.2150721Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-09-28T21:52:21.2151136Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-09-28T21:52:21.2151609Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-09-28T21:52:21.2152039Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-09-28T21:52:21.2152462Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-09-28T21:52:21.2152941Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-09-28T21:52:21.2153489Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-09-28T21:52:21.2153962Z  at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
> 2020-09-28T21:52:21.2154406Z  at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
> 2020-09-28T21:52:21.2154828Z  at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-09-28T21:52:21.2155381Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:367)
> 2020-09-28T21:52:21.2155864Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:274)
> 2020-09-28T21:52:21.2156378Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> 2020-09-28T21:52:21.2156865Z  at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:161)
> 2020-09-28T21:52:21.2157458Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:290)
> 2020-09-28T21:52:21.2157993Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:242)
> 2020-09-28T21:52:21.2158470Z  at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:121)
> 2020-09-28T21:52:21.2158890Z Caused by: java.lang.RuntimeException: Master 
> not initialized after 20ms
> 2020-09-28T21:52:21.2159350Z  at 
> org.apache.hadoop.hbase.util.JVMClusterUtil.waitForEvent(JVMClusterUtil.java:229)
> 2020-09-28T21:52:21.2159823Z  at 
> org.apache.hadoop.hbase.util.JVMClusterUtil.startup(JVMClusterUtil.java:197)
> 2020-09-28T21:52:21.2160270Z  at 
> org.apache.hadoop.hbase.LocalHBaseCluster.startup(LocalHBaseCluster.java:413)
> 2020-09-28T21:52:21.2160800Z  at 
> org.apache.hadoop.hbase.MiniHBaseCluster.init(MiniHBaseCluster.java:259)
> 2020-09-28T21:52:21.2161096Z  ... 22 more
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] dianfu closed pull request #13525: [FLINK-19447][hbase] Run HBase 2.2 integration test only if hadoop version is between 2.8.0 - 3.0.3

2020-10-01 Thread GitBox


dianfu closed pull request #13525:
URL: https://github.com/apache/flink/pull/13525


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19426) Streaming File Sink end-to-end test sometimes fails with "Could not assign resource ... to current execution ..."

2020-10-01 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205509#comment-17205509
 ] 

Dian Fu commented on FLINK-19426:
-

[~azagrebin] Thanks a lot for the investigation!

> Streaming File Sink end-to-end test sometimes fails with "Could not assign 
> resource ... to current execution ..."
> -
>
> Key: FLINK-19426
> URL: https://issues.apache.org/jira/browse/FLINK-19426
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network, Tests
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Assignee: Robert Metzger
>Priority: Major
>  Labels: pull-request-available, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=6983=logs=68a897ab-3047-5660-245a-cce8f83859f6=16ca2cca-2f63-5cce-12d2-d519b930a729
> {code}
> 2020-09-26T22:16:26.9856525Z 
> org.apache.flink.runtime.io.network.partition.consumer.PartitionConnectionException:
>  Connection for partition 
> 619775973ed0f282e20f9d55d13913ab#0@bc764cd8ddf7a0cff126f51c16239658_0_1 not 
> reachable.
> 2020-09-26T22:16:26.9857848Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.RemoteInputChannel.requestSubpartition(RemoteInputChannel.java:159)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9859168Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.internalRequestPartitions(SingleInputGate.java:336)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9860449Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.SingleInputGate.requestPartitions(SingleInputGate.java:308)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9861677Z  at 
> org.apache.flink.runtime.taskmanager.InputGateWithMetrics.requestPartitions(InputGateWithMetrics.java:95)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9862861Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.requestPartitions(StreamTask.java:542)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9864018Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.readRecoveredChannelState(StreamTask.java:507)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9865284Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.lambda$beforeInvoke$0(StreamTask.java:498)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9866415Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.runThrowing(StreamTaskActionExecutor.java:47)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9867500Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.beforeInvoke(StreamTask.java:492)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9868514Z  at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:550)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9869450Z  at 
> org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:722) 
> [flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9870339Z  at 
> org.apache.flink.runtime.taskmanager.Task.run(Task.java:547) 
> [flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9870869Z  at java.lang.Thread.run(Thread.java:748) 
> [?:1.8.0_265]
> 2020-09-26T22:16:26.9872060Z Caused by: java.io.IOException: 
> java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: 
> Connecting to remote task manager '/10.1.0.4:38905' has failed. This might 
> indicate that the remote task manager has been lost.
> 2020-09-26T22:16:26.9873511Z  at 
> org.apache.flink.runtime.io.network.netty.PartitionRequestClientFactory.createPartitionRequestClient(PartitionRequestClientFactory.java:85)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9874788Z  at 
> org.apache.flink.runtime.io.network.netty.NettyConnectionManager.createPartitionRequestClient(NettyConnectionManager.java:67)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9876084Z  at 
> org.apache.flink.runtime.io.network.partition.consumer.RemoteInputChannel.requestSubpartition(RemoteInputChannel.java:156)
>  ~[flink-dist_2.11-1.12-SNAPSHOT.jar:1.12-SNAPSHOT]
> 2020-09-26T22:16:26.9876567Z  ... 12 more
> 2020-09-26T22:16:26.9877477Z Caused by: 
> java.util.concurrent.ExecutionException: 
> org.apache.flink.runtime.io.network.netty.exception.RemoteTransportException: 
> Connecting to remote task manager '/10.1.0.4:38905' has failed. This might 
> indicate that the remote task manager has 

[jira] [Updated] (FLINK-19485) Decouple runtime operator implementation from DataStream operations

2020-10-01 Thread Aljoscha Krettek (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aljoscha Krettek updated FLINK-19485:
-
Summary: Decouple runtime operator implementation from DataStream 
operations   (was: Consider runtime-mode when creating the StreamGraph)

> Decouple runtime operator implementation from DataStream operations 
> 
>
> Key: FLINK-19485
> URL: https://issues.apache.org/jira/browse/FLINK-19485
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Affects Versions: 1.12.0
>Reporter: Kostas Kloudas
>Assignee: Kostas Kloudas
>Priority: Major
>
> Although DataStream is going to be the unified API for Batch and Streaming 
> applications, some operations, e.g. Sinks, may need to have different runtime 
> implementations depending on if they are intended to run on bounded or 
> unbounded data. This is not necessarily only for optimisations but also for 
> the exposed semantics, i.e. correctness.
> So far, DataStream had a 1-to-1 mapping between an API call and an operator. 
> In a sense, the DataStream API was an "explicit" API. With this addition, we 
> will decouple the API calls from the actual runtime implementations and thus 
> allow different operations to have more than one runtime implementations, 
> depending (for now) on the {{execution.runtime-mode}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] alpinegizmo commented on a change in pull request #13510: [FLINK-13095][state-processor-api] Introduce window bootstrap writer for writing window operator state

2020-10-01 Thread GitBox


alpinegizmo commented on a change in pull request #13510:
URL: https://github.com/apache/flink/pull/13510#discussion_r498222828



##
File path: docs/dev/libs/state_processor_api.md
##
@@ -742,6 +740,56 @@ If a processing time timer is set but the state is not 
restored until after that
 
 Attention If your bootstrap function 
creates timers, the state can only be restored using one of the [process]({{ 
site.baseurl }}/dev/stream/operators/process_function.html) type functions.
 
+### Window State
+
+The state processor api supports writing state for the [window operator]({{ 
site.baseurl }}/dev/stream/operators/windows.html).
+When writing window state, users specify the operator id, window assigner, 
evictor, optional trigger, and aggregation type.
+It is important the configurations on the bootstrap transformation match the 
configurations on the DataStream window.
+
+
+
+{% highlight java %}
+public class Account {
+public int id;
+
+public double amount;  
+
+public long timestamp;
+}
+ 
+ExecutionEnvironment bEnv = ExecutionEnvironment.getExecutionEnvironment();
+
+DataSet accountDataSet = bEnv.fromCollection(accounts);
+
+BootstrapTransformation transformation = OperatorTransformation
+.bootstrapWith(accountDataSet)
+// When using event time windows, its important

Review comment:
   ```suggestion
   // When using event time windows, it is important
   ```

##
File path: docs/dev/libs/state_processor_api.md
##
@@ -742,6 +740,56 @@ If a processing time timer is set but the state is not 
restored until after that
 
 Attention If your bootstrap function 
creates timers, the state can only be restored using one of the [process]({{ 
site.baseurl }}/dev/stream/operators/process_function.html) type functions.
 
+### Window State
+
+The state processor api supports writing state for the [window operator]({{ 
site.baseurl }}/dev/stream/operators/windows.html).
+When writing window state, users specify the operator id, window assigner, 
evictor, optional trigger, and aggregation type.
+It is important the configurations on the bootstrap transformation match the 
configurations on the DataStream window.
+
+
+
+{% highlight java %}
+public class Account {
+public int id;
+
+public double amount;  
+
+public long timestamp;
+}
+ 
+ExecutionEnvironment bEnv = ExecutionEnvironment.getExecutionEnvironment();
+
+DataSet accountDataSet = bEnv.fromCollection(accounts);
+
+BootstrapTransformation transformation = OperatorTransformation
+.bootstrapWith(accountDataSet)
+// When using event time windows, its important
+// to assign timestamps to each record.
+.assignTimestamps(account -> account.timestamp)
+.keyBy(acc -> acc.id)
+.window(TumblingEventTimeWindows.of)
+.reduce((left, right) -> left + right);
+{% endhighlight %}
+
+
+{% highlight scala %}
+case class Account(id: Int, amount: Double, timestamp: Long)
+ 
+val bEnv = ExecutionEnvironment.getExecutionEnvironment();
+val accountDataSet = bEnv.fromCollection(accounts);
+
+val transformation = OperatorTransformation
+.bootstrapWith(accountDataSet)
+// When using event time windows, its important
+// to assign timestamps to each record.
+.assignTimestamps(account -> account.timestamp)
+.keyBy(acc -> acc.id)
+.window(TumblingEventTimeWindows.of)
+.reduce((left, right) -> left + right)

Review comment:
   This doesn't look like code that would actually compile.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] dawidwys commented on a change in pull request #12056: [FLINK-17502] [flink-connector-rabbitmq] RMQSource refactor

2020-10-01 Thread GitBox


dawidwys commented on a change in pull request #12056:
URL: https://github.com/apache/flink/pull/12056#discussion_r498214070



##
File path: 
flink-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQDeserializationSchema.java
##
@@ -0,0 +1,91 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.connectors.rabbitmq;
+
+import org.apache.flink.api.common.serialization.DeserializationSchema;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.java.typeutils.ResultTypeQueryable;
+import org.apache.flink.util.Collector;
+
+import com.rabbitmq.client.AMQP;
+import com.rabbitmq.client.Envelope;
+
+import java.io.IOException;
+import java.io.Serializable;
+import java.util.List;
+
+/**
+ * Interface for the set of methods required to parse an RMQ delivery.
+ * @param  The output type of the {@link RMQSource}
+ */
+public interface RMQDeserializationSchema extends Serializable, 
ResultTypeQueryable {
+
+   /**
+* Initialization method for the schema. It is called before the actual 
working methods
+* {@link #deserialize} and thus suitable for one time setup work.
+*
+* The provided {@link DeserializationSchema.InitializationContext} 
can be used to access additional features such as e.g.
+* registering user metrics.
+*
+* @param context Contextual information that can be used during 
initialization.
+*/
+   public void open(DeserializationSchema.InitializationContext context) 
throws Exception;
+
+
+   /**
+* This method takes all the RabbitMQ delivery information supplied by 
the client extract the data and pass it to the
+* collector.
+* NOTICE: The implementation of this method MUST call {@link 
RMQCollector#setMessageIdentifiers(String, long)} with
+* the correlation ID of the message if checkpointing and 
UseCorrelationID (in the RMQSource constructor) were enabled
+* the {@link RMQSource}.
+* @param envelope
+* @param properties
+* @param body
+* @throws IOException
+*/

Review comment:
   ```suggestion
 * This method takes all the RabbitMQ delivery information supplied by 
the client extract the data and pass it to the
 * collector.
 *
 * NOTICE: The implementation of this method MUST call {@link 
RMQCollector#setMessageIdentifiers(String, long)} with
 * the correlation ID of the message if checkpointing and 
UseCorrelationID (in the RMQSource constructor) were enabled
 * the {@link RMQSource}.
 */
   ```

##
File path: 
flink-connectors/flink-connector-rabbitmq/src/main/java/org/apache/flink/streaming/connectors/rabbitmq/RMQDeserializationSchema.java
##
@@ -0,0 +1,91 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.connectors.rabbitmq;
+
+import org.apache.flink.api.common.serialization.DeserializationSchema;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.java.typeutils.ResultTypeQueryable;
+import org.apache.flink.util.Collector;
+
+import com.rabbitmq.client.AMQP;
+import com.rabbitmq.client.Envelope;
+
+import java.io.IOException;
+import java.io.Serializable;
+import java.util.List;
+
+/**
+ * Interface for the set of methods required to parse 

[GitHub] [flink] juha-mynttinen-king commented on pull request #13393: [FLINK-19238] [RocksDB] Sanity check for arena block size

2020-10-01 Thread GitBox


juha-mynttinen-king commented on pull request #13393:
URL: https://github.com/apache/flink/pull/13393#issuecomment-702109330


   Hey. I'm going to lose access to the GitHub account soon, and not be able to 
modify the branch anymore. If changes are needed before merge, I guess I could 
create a new PR using another GH account.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13521:
URL: https://github.com/apache/flink/pull/13521#issuecomment-701282459


   
   ## CI report:
   
   * ad55b5a30711435f6caf0d52ba1cdd702af96fda Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7151)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13525: [FLINK-19447][hbase] Run HBase 2.2 integration test only if hadoop version is between 2.8.0 - 3.0.3

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13525:
URL: https://github.com/apache/flink/pull/13525#issuecomment-701444731


   
   ## CI report:
   
   * df45481e9c3a07bcbfac6e070bc16a38fe83f3ce Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7147)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #13521:
URL: https://github.com/apache/flink/pull/13521#issuecomment-701282459


   
   ## CI report:
   
   * 4824a2e5258649246c5372200dd44fb61e20d012 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7143)
 
   * ad55b5a30711435f6caf0d52ba1cdd702af96fda UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18044) Add the subtask index information to the SourceReaderContext.

2020-10-01 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18044?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205439#comment-17205439
 ] 

Stephan Ewen commented on FLINK-18044:
--

I am not strictly opposed to exposing the Subtask Id.

It is more that this encourages users to write code that is not oblivious to 
parallelism, as code ideally should be.
Code that makes explicit assumptions about the current parallelism and the 
current subtask index would break when we try to introduce mode dynamic models 
(where some subtasks can split/scale out without affecting other tasks).

> Add the subtask index information to the SourceReaderContext.
> -
>
> Key: FLINK-18044
> URL: https://issues.apache.org/jira/browse/FLINK-18044
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Common
>Reporter: Jiangjie Qin
>Priority: Major
>  Labels: pull-request-available
>
> It is useful for the `SourceReader` to retrieve its subtask id. For example, 
> Kafka readers can create a consumer with proper client id.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] dawidwys commented on a change in pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


dawidwys commented on a change in pull request #13521:
URL: https://github.com/apache/flink/pull/13521#discussion_r498180993



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/CircularQueues.java
##
@@ -75,6 +75,9 @@ public void send(StageRunner.SortStage stage, 
CircularElement element) {
 
@Override
public void sendResult(MutableObjectIterator result) {
+   if (iteratorFuture.isDone()) {

Review comment:
   Actually these changes are unnecessary, I added them for helping me with 
debugging and forgot to remove them.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] dawidwys commented on a change in pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


dawidwys commented on a change in pull request #13521:
URL: https://github.com/apache/flink/pull/13521#discussion_r498177852



##
File path: 
flink-core/src/main/java/org/apache/flink/configuration/ExecutionOptions.java
##
@@ -55,4 +56,14 @@
"throughput")
)
.build());
+
+   @Documentation.ExcludeFromDocumentation("The option is considered 
internal and should rather not be used" +

Review comment:
   It won't work in `StreamNode`. Only `StreamConfig` is shipped to 
`StreamTask`.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] zentol commented on pull request #384: Improved Docs

2020-10-01 Thread GitBox


zentol commented on pull request #384:
URL: https://github.com/apache/flink-web/pull/384#issuecomment-702075483


   I do not think this really enhances the documentation.
   
   If you wish to make a contribution to Flink, please read the [contribution 
guide](https://flink.apache.org/contributing/how-to-contribute.html). I'm sure 
you can find something else to work on.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] zentol closed pull request #384: Improved Docs

2020-10-01 Thread GitBox


zentol closed pull request #384:
URL: https://github.com/apache/flink-web/pull/384


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19441) Performance regression on 24.09.2020

2020-10-01 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205424#comment-17205424
 ] 

Stephan Ewen commented on FLINK-19441:
--

[~AHeise] That sounds like a plan. I'll start digging into the code, see if I 
can spot some virtual per-record calls.

> Performance regression on 24.09.2020
> 
>
> Key: FLINK-19441
> URL: https://issues.apache.org/jira/browse/FLINK-19441
> Project: Flink
>  Issue Type: Bug
>Reporter: Arvid Heise
>Priority: Blocker
>
> A couple of benchmarks are showing a small performance regression on 
> 24.09.2020:
> http://codespeed.dak8s.net:8000/timeline/?ben=globalWindow=2
> http://codespeed.dak8s.net:8000/timeline/?ben=tupleKeyBy=2 (?)



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] StephanEwen commented on pull request #13512: [FLINK-19457][core] Add a number sequence generating source for the New Source API.

2020-10-01 Thread GitBox


StephanEwen commented on pull request #13512:
URL: https://github.com/apache/flink/pull/13512#issuecomment-702065085


   For both topics, this PR here stays close to what the previous state of 
sources was, in an effort to port the previous state to the new API.
   
   We can open the discussions about having different package structures in the 
future, but do we need to block this PR on that?
   The package structure in flink-core has never separated API and 
implementation, and the `api` name was not chosen with very much consideration. 
IIRC mainly to separate SDK from runtime parts. (Arguable `.sdk.` would have 
been the correct name, but were were all younger and less experienced when we 
created these parts.)
   
   From the user's perspective I think it is better to stay consistent within 
one style then have different packages follow different styles. That adds more 
confusion and makes it harder to navigate.
   
   Similar for the "deep integration", I feel this should not block this PR 
(foremost because this PR does not touch anything there). 
   That is a different discussion that we should either apply everywhere or not 
do it. Doing this now here differently than in all other places will add 
confusion for users. Another side of the problem is that many of these methods 
`env.generateSequence(...)` are public API that we should not break.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-19479) Allow explicitly configuring time behaviour on KeyedStream.intervalJoin()

2020-10-01 Thread Nicholas Jiang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205423#comment-17205423
 ] 

Nicholas Jiang commented on FLINK-19479:


[~aljoscha], IMO, I could also work for this issue.:D

> Allow explicitly configuring time behaviour on KeyedStream.intervalJoin()
> -
>
> Key: FLINK-19479
> URL: https://issues.apache.org/jira/browse/FLINK-19479
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Aljoscha Krettek
>Priority: Major
>
> With the deprecation of 
> {{StreamExecutionEnvironment.setStreamTimeCharacteristic()}} in FLINK-19319 
> we need a way of explicitly configuring the time behaviour of these join 
> operations. Currently, all join operations use the characteristic to 
> configure themselves.
> Alternatively, we might consider removing/deprecating these join operations.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19249) Detect broken connections in case TCP Timeout takes too long.

2020-10-01 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-19249:
-
Summary: Detect broken connections in case TCP Timeout takes too long.  
(was: Job would wait sometime(~10 min) before failover if some connection 
broken)

> Detect broken connections in case TCP Timeout takes too long.
> -
>
> Key: FLINK-19249
> URL: https://issues.apache.org/jira/browse/FLINK-19249
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Congxian Qiu(klion26)
>Priority: Major
>
> {quote}encountered this error on 1.7, after going through the master code, I 
> think the problem is still there
> {quote}
> When the network environment is not so good, the connection between the 
> server and the client may be disconnected innocently. After the 
> disconnection, the server will receive the IOException such as below
> {code:java}
> java.io.IOException: Connection timed out
>  at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
>  at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
>  at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
>  at sun.nio.ch.IOUtil.write(IOUtil.java:51)
>  at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:468)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doWrite(NioSocketChannel.java:403)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.flush0(AbstractChannel.java:934)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.forceFlush(AbstractNioChannel.java:367)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:639)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
>  at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884)
>  at java.lang.Thread.run(Thread.java:748)
> {code}
> then release the view reader.
> But the job would not fail until the downstream detect the disconnection 
> because of {{channelInactive}} later(~10 min). between such time, the job can 
> still process data, but the broken channel can't transfer any data or event, 
> so snapshot would fail during this time. this will cause the job to replay 
> many data after failover.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19249) Job would wait sometime(~10 min) before failover if some connection broken

2020-10-01 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-19249:
-
Priority: Major  (was: Critical)

> Job would wait sometime(~10 min) before failover if some connection broken
> --
>
> Key: FLINK-19249
> URL: https://issues.apache.org/jira/browse/FLINK-19249
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Congxian Qiu(klion26)
>Priority: Major
>
> {quote}encountered this error on 1.7, after going through the master code, I 
> think the problem is still there
> {quote}
> When the network environment is not so good, the connection between the 
> server and the client may be disconnected innocently. After the 
> disconnection, the server will receive the IOException such as below
> {code:java}
> java.io.IOException: Connection timed out
>  at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
>  at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
>  at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
>  at sun.nio.ch.IOUtil.write(IOUtil.java:51)
>  at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:468)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doWrite(NioSocketChannel.java:403)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.flush0(AbstractChannel.java:934)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.forceFlush(AbstractNioChannel.java:367)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:639)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
>  at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884)
>  at java.lang.Thread.run(Thread.java:748)
> {code}
> then release the view reader.
> But the job would not fail until the downstream detect the disconnection 
> because of {{channelInactive}} later(~10 min). between such time, the job can 
> still process data, but the broken channel can't transfer any data or event, 
> so snapshot would fail during this time. this will cause the job to replay 
> many data after failover.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19249) Job would wait sometime(~10 min) before failover if some connection broken

2020-10-01 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17205414#comment-17205414
 ] 

Stephan Ewen commented on FLINK-19249:
--

Digging into this a bit more, I think your are right. This is a corner case 
caused by problematic network environments.

Let's label this as an improvement where Flink could try to do something on the 
application layer to detect network issues when the kernel takes too long.

But we still miss a good proposal how to do that.

> Job would wait sometime(~10 min) before failover if some connection broken
> --
>
> Key: FLINK-19249
> URL: https://issues.apache.org/jira/browse/FLINK-19249
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Congxian Qiu(klion26)
>Priority: Critical
> Fix For: 1.12.0
>
>
> {quote}encountered this error on 1.7, after going through the master code, I 
> think the problem is still there
> {quote}
> When the network environment is not so good, the connection between the 
> server and the client may be disconnected innocently. After the 
> disconnection, the server will receive the IOException such as below
> {code:java}
> java.io.IOException: Connection timed out
>  at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
>  at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
>  at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
>  at sun.nio.ch.IOUtil.write(IOUtil.java:51)
>  at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:468)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doWrite(NioSocketChannel.java:403)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.flush0(AbstractChannel.java:934)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.forceFlush(AbstractNioChannel.java:367)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:639)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
>  at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884)
>  at java.lang.Thread.run(Thread.java:748)
> {code}
> then release the view reader.
> But the job would not fail until the downstream detect the disconnection 
> because of {{channelInactive}} later(~10 min). between such time, the job can 
> still process data, but the broken channel can't transfer any data or event, 
> so snapshot would fail during this time. this will cause the job to replay 
> many data after failover.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-19249) Job would wait sometime(~10 min) before failover if some connection broken

2020-10-01 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-19249:
-
Fix Version/s: (was: 1.12.0)

> Job would wait sometime(~10 min) before failover if some connection broken
> --
>
> Key: FLINK-19249
> URL: https://issues.apache.org/jira/browse/FLINK-19249
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Network
>Reporter: Congxian Qiu(klion26)
>Priority: Critical
>
> {quote}encountered this error on 1.7, after going through the master code, I 
> think the problem is still there
> {quote}
> When the network environment is not so good, the connection between the 
> server and the client may be disconnected innocently. After the 
> disconnection, the server will receive the IOException such as below
> {code:java}
> java.io.IOException: Connection timed out
>  at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
>  at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
>  at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
>  at sun.nio.ch.IOUtil.write(IOUtil.java:51)
>  at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:468)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.socket.nio.NioSocketChannel.doWrite(NioSocketChannel.java:403)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.AbstractChannel$AbstractUnsafe.flush0(AbstractChannel.java:934)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.forceFlush(AbstractNioChannel.java:367)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:639)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:580)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:497)
>  at 
> org.apache.flink.shaded.netty4.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:459)
>  at 
> org.apache.flink.shaded.netty4.io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:884)
>  at java.lang.Thread.run(Thread.java:748)
> {code}
> then release the view reader.
> But the job would not fail until the downstream detect the disconnection 
> because of {{channelInactive}} later(~10 min). between such time, the job can 
> still process data, but the broken channel can't transfer any data or event, 
> so snapshot would fail during this time. this will cause the job to replay 
> many data after failover.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12056: [FLINK-17502] [flink-connector-rabbitmq] RMQSource refactor

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #12056:
URL: https://github.com/apache/flink/pull/12056#issuecomment-626167900


   
   ## CI report:
   
   * c42ef1f62396948ece865a3b8628bea229d5ffb9 UNKNOWN
   * ad971fd47ab3ac9e813afae9713c75986646 UNKNOWN
   * 4d2d57b8c2f6810c03d284fdd5807b81a2b554d7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7120)
 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7150)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7142)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-playgrounds] alpinegizmo commented on a change in pull request #16: [FLINK-19145][walkthroughs] Add PyFlink-walkthrough to Flink playground.

2020-10-01 Thread GitBox


alpinegizmo commented on a change in pull request #16:
URL: https://github.com/apache/flink-playgrounds/pull/16#discussion_r498114079



##
File path: pyflink-walkthrough/README.md
##
@@ -0,0 +1,140 @@
+# pyflink-walkthrough
+
+## Background
+
+In this playground, you will learn how to build and run an end-to-end PyFlink 
pipeline for data analytics, covering the following steps:
+
+* Reading data from a Kafka source;
+* Creating data using a 
[UDF](https://ci.apache.org/projects/flink/flink-docs-release-1.11/dev/python/table-api-users-guide/udfs/python_udfs.html);
+* Performing a simple aggregation over the source data;
+* Writing the results to Elasticsearch and visualizing them in Kibana.
+
+The environment is based on Docker Compose, so the only requirement is that 
you have [Docker](https://docs.docker.com/get-docker/) 
+installed in your machine.
+
+### Kafka
+You will be using Kafka to store sample input data about payment transactions. 
A simple data generator 
[generate_source_data.py](generator/generate_source_data.py) is provided to
+continuously write new records to the `payment_msg` Kafka topic. Each record 
is structured as follows:
+ 
+`{"createTime": "2020-08-12 06:29:02", "orderId": 1597213797, "payAmount": 
28306.44976403719, "payPlatform": 0, "provinceId": 4}`
+
+* `createTime`: The creation time of the transaction. 
+* `orderId`: The id of the current transaction.
+* `payAmount`: The amount being paid with this transaction.
+* `payPlatform`: The platform used to create this payment: pc or mobile.
+* `provinceId`: The id of the province for the user. 
+
+You can use the following command to read data from the Kafka topic and check 
whether it's generated correctly:
+```shell script
+$ docker-compose exec kafka kafka-console-consumer.sh --bootstrap-server 
kafka:9092 --topic payment_msg
+{"createTime":"2020-07-27 
09:25:32.77","orderId":1595841867217,"payAmount":7732.44,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.231","orderId":1595841867218,"payAmount":75774.05,"payPlatform":0,"provinceId":3}
+{"createTime":"2020-07-27 
09:25:33.72","orderId":1595841867219,"payAmount":65908.55,"payPlatform":0,"provinceId":0}
+{"createTime":"2020-07-27 
09:25:34.216","orderId":1595841867220,"payAmount":15341.11,"payPlatform":0,"provinceId":1}
+{"createTime":"2020-07-27 
09:25:34.698","orderId":1595841867221,"payAmount":37504.42,"payPlatform":0,"provinceId":0}
+```
+You can also create a new topic by executing the following command:
+```shell script
+$ docker-compose exec kafka kafka-topics.sh --bootstrap-server kafka:9092 
--create --topic  --partitions 8 --replication-factor 1
+```
+
+### PyFlink
+
+The transaction data will be processed with PyFlink using the Python script 
[payment_msg_processing.py](payment_msg_proccessing.py).
+This script will first map the `provinceId` in the input records to its 
corresponding province name using a Python UDF, 
+and then compute the sum of the transaction amounts for each province.
+
+### ElasticSearch
+
+ElasticSearch is used to store the results and to provide an efficient query 
service.
+
+### Kibana
+
+Kibana is an open source data visualization dashboard for ElasticSearch. You 
will use it to visualize 
+the results of your PyFlink pipeline.
+
+## Setup
+
+As mentioned, the environment for this walkthrough is based on Docker Compose; 
It uses a custom image
+to spin up Flink (JobManager+TaskManager), Kafka+Zookeeper, the data 
generator, and Elasticsearch+Kibana containers.
+
+You can find the [docker-compose.yaml](docker-compose.yml) file of the 
pyflink-walkthrough in the `pyflink-walkthrough` root directory.
+
+### Building the Docker image
+
+First, build the Docker image by running:
+
+```bash
+docker-compose build
+```
+
+### Starting the Playground
+
+Once the Docker image build is complete, run the following command to start 
the playground:
+
+```bash
+docker-compose up -d
+```
+
+One way of checking if the playground was successfully started is to access 
some of the services that are exposed:
+
+1. visiting Flink Web UI [http://localhost:8081](http://localhost:8081).
+2. visiting Elasticsearch [http://localhost:9200](http://localhost:9200).
+3. visiting Kibana [http://localhost:5601](http://localhost:5601).
+
+**Note:** you may need to wait around 1 minute before all the services come up.
+
+### Stopping the Playground
+
+To stop the playground, run the following command:
+
+```bash
+docker-compose down
+```
+
+
+## Running the PyFlink job
+
+1. Submit the PyFlink job.
+```shell script
+$ docker-compose exec jobmanager ./bin/flink run -py 
/opt/pyflink-walkthrough/payment_msg_proccessing.py -d
+```
+Navigate to the [Flink Web UI](http://localhost:8081) after the job is 
submitted successfully. There should be a job in the running job list.
+Click the job to get more details. You should see that the `StreamGraph` of 
the `payment_msg_proccessing` consists of two nodes, each with a parallelism of 
1. 

[GitHub] [flink] aljoscha commented on a change in pull request #13521: [FLINK-19472] Implement a one input sorting DataInput

2020-10-01 Thread GitBox


aljoscha commented on a change in pull request #13521:
URL: https://github.com/apache/flink/pull/13521#discussion_r498141996



##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/sort/KeyAndValueSerializer.java
##
@@ -0,0 +1,181 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.operators.sort;
+
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.common.typeutils.TypeSerializerSnapshot;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.core.memory.DataInputView;
+import org.apache.flink.core.memory.DataOutputView;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+
+import java.io.IOException;
+import java.util.Arrays;
+import java.util.Objects;
+
+/**
+ * A serializer used in {@link SortingDataInput} for serializing elements 
alongside their key and
+ * timestamp. It serializes the record in a format known by the {@link 
FixedLengthByteKeyComparator}
+ * and {@link VariableLengthByteKeyComparator}.
+ *
+ * If the key is of known constant length, the length is not serialized 
with the data.
+ * Therefore the serialized data is as follows:
+ *
+ * 
+ *  [key-length] | key | timestamp | record
+ * 
+ */
+final class KeyAndValueSerializer extends TypeSerializer>> {
+   private final TypeSerializer valueSerializer;
+   private final int serializedKeyLength;

Review comment:
   Maybe add a comment explaining that this is either a fixed key length or 
negative if it is variable.

##
File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/sort/SortingDataInput.java
##
@@ -0,0 +1,227 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.operators.sort;
+
+import org.apache.flink.api.common.typeutils.TypeComparator;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+import org.apache.flink.api.java.functions.KeySelector;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.configuration.AlgorithmOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.core.io.InputStatus;
+import org.apache.flink.core.memory.DataOutputSerializer;
+import org.apache.flink.runtime.checkpoint.channel.ChannelStateWriter;
+import org.apache.flink.runtime.io.AvailabilityProvider;
+import org.apache.flink.runtime.io.disk.iomanager.IOManager;
+import org.apache.flink.runtime.jobgraph.tasks.AbstractInvokable;
+import org.apache.flink.runtime.memory.MemoryAllocationException;
+import org.apache.flink.runtime.memory.MemoryManager;
+import org.apache.flink.runtime.operators.sort.ExternalSorter;
+import org.apache.flink.runtime.operators.sort.PushSorter;
+import org.apache.flink.streaming.api.watermark.Watermark;
+import org.apache.flink.streaming.runtime.io.StreamMultipleInputProcessor;
+import org.apache.flink.streaming.runtime.io.StreamTaskInput;
+import org.apache.flink.streaming.runtime.streamrecord.LatencyMarker;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+import org.apache.flink.streaming.runtime.streamstatus.StreamStatus;
+import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.MutableObjectIterator;
+
+import javax.annotation.Nonnull;
+
+import java.io.IOException;
+import java.util.concurrent.CompletableFuture;
+
+/**
+ * A 

[GitHub] [flink] flinkbot edited a comment on pull request #12056: [FLINK-17502] [flink-connector-rabbitmq] RMQSource refactor

2020-10-01 Thread GitBox


flinkbot edited a comment on pull request #12056:
URL: https://github.com/apache/flink/pull/12056#issuecomment-626167900


   
   ## CI report:
   
   * c42ef1f62396948ece865a3b8628bea229d5ffb9 UNKNOWN
   * ad971fd47ab3ac9e813afae9713c75986646 UNKNOWN
   * 4d2d57b8c2f6810c03d284fdd5807b81a2b554d7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7142)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7120)
 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=7150)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] senegalo commented on pull request #12056: [FLINK-17502] [flink-connector-rabbitmq] RMQSource refactor

2020-10-01 Thread GitBox


senegalo commented on pull request #12056:
URL: https://github.com/apache/flink/pull/12056#issuecomment-702037986


   @flinkbot run azure



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




  1   2   >