[jira] [Updated] (FLINK-33463) Support the implementation of dynamic source tables based on the new source

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-33463:
---
Labels: pull-request-available  (was: )

> Support the implementation of dynamic source tables based on the new source
> ---
>
> Key: FLINK-33463
> URL: https://issues.apache.org/jira/browse/FLINK-33463
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / JDBC
>Reporter: RocMarshal
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-33463][Connector/JDBC] Support the implementation of dynamic source tables based on the new source [flink-connector-jdbc]

2024-04-19 Thread via GitHub


RocMarshal opened a new pull request, #117:
URL: https://github.com/apache/flink-connector-jdbc/pull/117

   - Support the implementation of dynamic source tables based on the new source


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33379) Bump CI flink version on flink-connector-elasticsearch

2024-04-19 Thread Yubin Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33379?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17839184#comment-17839184
 ] 

Yubin Li commented on FLINK-33379:
--

[~pgodowski] sorry for the late reply, it seems that 3.1.0 branch not created 
yet, you could compile using master branch to support Flink 1.18

> Bump CI flink version on flink-connector-elasticsearch
> --
>
> Key: FLINK-33379
> URL: https://issues.apache.org/jira/browse/FLINK-33379
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: elasticsearch-3.1.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>  Labels: pull-request-available
> Fix For: elasticsearch-3.1.0
>
> Attachments: image-2023-10-27-16-54-04-937.png
>
>
> As Flink 1.18 released, bump the flink version in es connector .



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34918) Introduce comment for Catalog

2024-04-19 Thread Yubin Li (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yubin Li updated FLINK-34918:
-
Summary: Introduce comment for Catalog  (was: Introduce comment support for 
Catalog)

> Introduce comment for Catalog
> -
>
> Key: FLINK-34918
> URL: https://issues.apache.org/jira/browse/FLINK-34918
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.20.0
>Reporter: Yubin Li
>Assignee: Yubin Li
>Priority: Major
>
> We propose to introduce `getComment()` method in `CatalogDescriptor`, and the 
> reasons are as follows.
> 1. For the sake of design consistency, follow the design of FLIP-295 [1] 
> which introduced `CatalogStore` component, `CatalogDescriptor` includes names 
> and attributes, both of which are used to describe the catalog, and `comment` 
> can be added smoothly.
> 2. Extending the existing class rather than add new method to the existing 
> interface, Especially, the `Catalog` interface, as a core interface, is used 
> by a series of important components such as `CatalogFactory`, 
> `CatalogManager` and `FactoryUtil`, and is implemented by a large number of 
> connectors such as JDBC, Paimon, and Hive. Adding methods to it will greatly 
> increase the implementation complexity, and more importantly, increase the 
> cost of iteration, maintenance, and verification.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35177] Fix DataGen Connector documentation [flink]

2024-04-19 Thread via GitHub


flinkbot commented on PR #24692:
URL: https://github.com/apache/flink/pull/24692#issuecomment-2067479443

   
   ## CI report:
   
   * 4d0723dfc64326e5fab53b2b59c495f341a865df UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35177) Datagen examples in documentation do not compile

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35177?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35177:
---
Labels: pull-request-available  (was: )

> Datagen examples in documentation do not compile
> 
>
> Key: FLINK-35177
> URL: https://issues.apache.org/jira/browse/FLINK-35177
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Sergei Morozov
>Priority: Not a Priority
>  Labels: pull-request-available
>
> Currently, the 
> [examples|https://nightlies.apache.org/flink/flink-docs-release-1.19/docs/connectors/datastream/datagen/#rate-limiting]
>  look like this:
> {code:java}
> GeneratorFunction generatorFunction = index -> index;
> double recordsPerSecond = 100;
> DataGeneratorSource source =
> new DataGeneratorSource<>(
>  generatorFunction,
>  Long.MAX_VALUE,
>  RateLimiterStrategy.perSecond(recordsPerSecond),
>  Types.STRING);
> {code}
> The generator function returns Long but the DataGeneratorSource uses String, 
> so their types do not match.
> Either the generator function needs to be changed to return a string, or the 
> source needs to use Long.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-35177] Fix DataGen Connector documentation [flink]

2024-04-19 Thread via GitHub


morozov opened a new pull request, #24692:
URL: https://github.com/apache/flink/pull/24692

   
   
   ## What is the purpose of the change
   
   The code 
[examples](https://nightlies.apache.org/flink/flink-docs-release-1.19/docs/connectors/datastream/datagen/#rate-limiting)
 used in the documentation are incorrect. Currently, they look like this:
   ```java
   GeneratorFunction generatorFunction = index -> index;
   double recordsPerSecond = 100;
   
   DataGeneratorSource source =
   new DataGeneratorSource<>(
generatorFunction,
Long.MAX_VALUE,
RateLimiterStrategy.perSecond(recordsPerSecond),
Types.STRING);
   ```
   
   The generator function returns `Long `but the `DataGeneratorSource` uses 
`String` and `Types.STRING`, so the types do not match, and the example code 
cannot be compiled.
   
   ## Brief change log
   
   The types used by the `DataGeneratorSource` are updated to match the return 
type of the generator function.
   
   ## Verifying this change
   
   Please make sure both new and modified tests in this PR follow [the 
conventions for tests defined in our code quality 
guide](https://flink.apache.org/how-to-contribute/code-style-and-quality-common/#7-testing).
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35177) Datagen examples in documentation do not compile

2024-04-19 Thread Sergei Morozov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35177?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergei Morozov updated FLINK-35177:
---
Description: 
Currently, the 
[examples|https://nightlies.apache.org/flink/flink-docs-release-1.19/docs/connectors/datastream/datagen/#rate-limiting]
 look like this:
{code:java}
GeneratorFunction generatorFunction = index -> index;
double recordsPerSecond = 100;

DataGeneratorSource source =
new DataGeneratorSource<>(
 generatorFunction,
 Long.MAX_VALUE,
 RateLimiterStrategy.perSecond(recordsPerSecond),
 Types.STRING);
{code}
The generator function returns Long but the DataGeneratorSource uses String, so 
their types do not match.

Either the generator function needs to be changed to return a string, or the 
source needs to use Long.

  was:
Currently, the examples look like this:

{code}
GeneratorFunction generatorFunction = index -> index;
double recordsPerSecond = 100;

DataGeneratorSource source =
new DataGeneratorSource<>(
 generatorFunction,
 Long.MAX_VALUE,
 RateLimiterStrategy.perSecond(recordsPerSecond),
 Types.STRING);
{code}

The generator function returns Long but the DataGeneratorSource uses String, so 
their types do not match.

Either the generator function needs to be changed to return a string, or the 
source needs to use Long.


> Datagen examples in documentation do not compile
> 
>
> Key: FLINK-35177
> URL: https://issues.apache.org/jira/browse/FLINK-35177
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.19.0
>Reporter: Sergei Morozov
>Priority: Not a Priority
>
> Currently, the 
> [examples|https://nightlies.apache.org/flink/flink-docs-release-1.19/docs/connectors/datastream/datagen/#rate-limiting]
>  look like this:
> {code:java}
> GeneratorFunction generatorFunction = index -> index;
> double recordsPerSecond = 100;
> DataGeneratorSource source =
> new DataGeneratorSource<>(
>  generatorFunction,
>  Long.MAX_VALUE,
>  RateLimiterStrategy.perSecond(recordsPerSecond),
>  Types.STRING);
> {code}
> The generator function returns Long but the DataGeneratorSource uses String, 
> so their types do not match.
> Either the generator function needs to be changed to return a string, or the 
> source needs to use Long.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35177) Datagen examples in documentation do not compile

2024-04-19 Thread Sergei Morozov (Jira)
Sergei Morozov created FLINK-35177:
--

 Summary: Datagen examples in documentation do not compile
 Key: FLINK-35177
 URL: https://issues.apache.org/jira/browse/FLINK-35177
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Affects Versions: 1.19.0
Reporter: Sergei Morozov


Currently, the examples look like this:

{code}
GeneratorFunction generatorFunction = index -> index;
double recordsPerSecond = 100;

DataGeneratorSource source =
new DataGeneratorSource<>(
 generatorFunction,
 Long.MAX_VALUE,
 RateLimiterStrategy.perSecond(recordsPerSecond),
 Types.STRING);
{code}

The generator function returns Long but the DataGeneratorSource uses String, so 
their types do not match.

Either the generator function needs to be changed to return a string, or the 
source needs to use Long.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35175][filesystem] Avoid reading with ByteBuffer on hadoop version below 3.3.0 [flink]

2024-04-19 Thread via GitHub


masteryhx commented on PR #24691:
URL: https://github.com/apache/flink/pull/24691#issuecomment-2066998011

   @Zakelly @RyanSkraba Could you help to take a look ?
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35175) HadoopDataInputStream can't compile with Hadoop 3.2.3

2024-04-19 Thread Hangxiang Yu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17839068#comment-17839068
 ] 

Hangxiang Yu commented on FLINK-35175:
--

[~rskraba] Thanks for reporting this!

I think it's better if we could upgrade the hadoop version, and I also saw 
before ticket about it: https://issues.apache.org/jira/browse/FLINK-33584

But seems there still are some blockers.

So I'd prefer to prepare a PR to remove the optimization currently and add TODO 
here. 

I will also try to help to push forward the upgradation.
Thanks again.

> HadoopDataInputStream can't compile with Hadoop 3.2.3
> -
>
> Key: FLINK-35175
> URL: https://issues.apache.org/jira/browse/FLINK-35175
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.20.0
>Reporter: Ryan Skraba
>Priority: Critical
>  Labels: pull-request-available
>
> Unfortunately, introduced in FLINK-35045: 
> [PREADWRITEBUFFER|https://github.com/apache/flink/commit/a312a3bdd258e0ff7d6f94e979b32e2bc762b82f#diff-3ed57be01895ba0f792110e40f4283427c55528f11a5105b4bf34ebd4e6fef0dR182]
>  was added in Hadoop releases 
> [3.3.0|https://github.com/apache/hadoop/blob/rel/release-3.3.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72]
>  and 
> [2.10.0|https://github.com/apache/hadoop/blob/rel/release-2.10.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72].
> It doesn't exist in flink.hadoop.version 
> [3.2.3|https://github.com/apache/hadoop/blob/rel/release-3.2.3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java],
>  which we are using in end-to-end tests.
> {code:java}
> 00:23:55.093 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.8.0:compile 
> (default-compile) on project flink-hadoop-fs: Compilation failure: 
> Compilation failure: 
> 00:23:55.093 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[151,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable READBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[182,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable PREADBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[183,43]
>  incompatible types: long cannot be converted to 
> org.apache.hadoop.io.ByteBufferPool
> 00:23:55.094 [ERROR] -> [Help 1] {code}
> * 1.20 compile_cron_hadoop313 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35175][filesystem] Avoid reading with ByteBuffer on hadoop version below 3.3.0 [flink]

2024-04-19 Thread via GitHub


flinkbot commented on PR #24691:
URL: https://github.com/apache/flink/pull/24691#issuecomment-2066992465

   
   ## CI report:
   
   * 38765977f787a0f88f247cd9ccfa7660f1c4d80b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35175) HadoopDataInputStream can't compile with Hadoop 3.2.3

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35175?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35175:
---
Labels: pull-request-available  (was: )

> HadoopDataInputStream can't compile with Hadoop 3.2.3
> -
>
> Key: FLINK-35175
> URL: https://issues.apache.org/jira/browse/FLINK-35175
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.20.0
>Reporter: Ryan Skraba
>Priority: Critical
>  Labels: pull-request-available
>
> Unfortunately, introduced in FLINK-35045: 
> [PREADWRITEBUFFER|https://github.com/apache/flink/commit/a312a3bdd258e0ff7d6f94e979b32e2bc762b82f#diff-3ed57be01895ba0f792110e40f4283427c55528f11a5105b4bf34ebd4e6fef0dR182]
>  was added in Hadoop releases 
> [3.3.0|https://github.com/apache/hadoop/blob/rel/release-3.3.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72]
>  and 
> [2.10.0|https://github.com/apache/hadoop/blob/rel/release-2.10.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72].
> It doesn't exist in flink.hadoop.version 
> [3.2.3|https://github.com/apache/hadoop/blob/rel/release-3.2.3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java],
>  which we are using in end-to-end tests.
> {code:java}
> 00:23:55.093 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.8.0:compile 
> (default-compile) on project flink-hadoop-fs: Compilation failure: 
> Compilation failure: 
> 00:23:55.093 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[151,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable READBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[182,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable PREADBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[183,43]
>  incompatible types: long cannot be converted to 
> org.apache.hadoop.io.ByteBufferPool
> 00:23:55.094 [ERROR] -> [Help 1] {code}
> * 1.20 compile_cron_hadoop313 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-35175][filesystem] Avoid reading with ByteBuffer on hadoop version below 3.3.0 [flink]

2024-04-19 Thread via GitHub


masteryhx opened a new pull request, #24691:
URL: https://github.com/apache/flink/pull/24691

   
   
   ## What is the purpose of the change
   
   Hadoop whose version below 3.3.0 doesn't support reading and positionable 
reading  with ByteBuffer, which may cause compile error. 
   See 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630
   
   
   ## Brief change log
   
   - Remove support about reading and positionable reading  with ByteBuffer, 
and add TODO.
   
   
   ## Verifying this change
   
   
   This change could be verified by current CI.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't 
know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-28048][connectors] Introduce Source API alternative to FiniteTestSource [flink]

2024-04-19 Thread via GitHub


tweise merged PR #23777:
URL: https://github.com/apache/flink/pull/23777


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-33460][Connector/JDBC] Support property authentication connection for JDBC catalog & dynamic table [flink-connector-jdbc]

2024-04-19 Thread via GitHub


RocMarshal opened a new pull request, #116:
URL: https://github.com/apache/flink-connector-jdbc/pull/116

   - Support property authentication connection for JDBC catalog & dynamic table


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-34574) Add CPU and memory size autoscaler quota

2024-04-19 Thread Gyula Fora (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34574?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyula Fora closed FLINK-34574.
--
Fix Version/s: kubernetes-operator-1.9.0
   Resolution: Fixed

merged to main baad90088ea5b5b240186a530a79b64fb84cc77e

> Add CPU and memory size autoscaler quota
> 
>
> Key: FLINK-34574
> URL: https://issues.apache.org/jira/browse/FLINK-34574
> Project: Flink
>  Issue Type: New Feature
>  Components: Autoscaler
>Reporter: Gabor Somogyi
>Assignee: Gabor Somogyi
>Priority: Major
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.9.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34574] Add CPU and memory size autoscaler quota [flink-kubernetes-operator]

2024-04-19 Thread via GitHub


gyfora merged PR #789:
URL: https://github.com/apache/flink-kubernetes-operator/pull/789


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35176) Support property authentication connection for JDBC catalog & dynamic table

2024-04-19 Thread RocMarshal (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35176?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

RocMarshal updated FLINK-35176:
---
Parent: FLINK-25420
Issue Type: Sub-task  (was: Improvement)

> Support property authentication connection for JDBC catalog & dynamic table
> ---
>
> Key: FLINK-35176
> URL: https://issues.apache.org/jira/browse/FLINK-35176
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC
>Reporter: RocMarshal
>Priority: Minor
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35176) Support property authentication connection for JDBC catalog & dynamic table

2024-04-19 Thread RocMarshal (Jira)
RocMarshal created FLINK-35176:
--

 Summary: Support property authentication connection for JDBC 
catalog & dynamic table
 Key: FLINK-35176
 URL: https://issues.apache.org/jira/browse/FLINK-35176
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / JDBC
Reporter: RocMarshal






--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-35159) CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash

2024-04-19 Thread Chesnay Schepler (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chesnay Schepler closed FLINK-35159.

Resolution: Fixed

> CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash
> 
>
> Key: FLINK-35159
> URL: https://issues.apache.org/jira/browse/FLINK-35159
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.18.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.2, 1.20.0, 1.19.1
>
>
> When a task manager dies while the JM is generating an ExecutionGraph in the 
> background then {{CreatingExecutionGraph#handleExecutionGraphCreation}} can 
> transition back into WaitingForResources if the TM hosted one of the slots 
> that we planned to use in {{tryToAssignSlots}}.
> At this point the ExecutionGraph was already transitioned to running, which 
> implicitly kicks of periodic checkpointing by the CheckpointCoordinator, 
> without the operator coordinator holders being initialized yet (as this 
> happens after we assigned slots).
> This effectively leaks that CheckpointCoordinator, including the timer thread 
> that will continue to try triggering checkpoints, which will naturally fail 
> to trigger.
> This can cause a JM crash because it results in 
> {{OperatorCoordinatorHolder#abortCurrentTriggering}} to be called, which 
> fails with an NPE since the {{mainThreadExecutor}} was not initialized yet.
> {code}
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$8(CheckpointCoordinator.java:707)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:610)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:910)
>   at 
> java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478)
>   at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>   at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>   at 
> java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>   at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.util.concurrent.CompletionException: 
> java.lang.NullPointerException
>   at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
>   at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
>   ... 7 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.abortCurrentTriggering(OperatorCoordinatorHolder.java:388)
>   at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
>   at 
> java.base/java.util.Collections$UnmodifiableCollection.forEach(Collections.java:1085)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:985)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:961)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$7(CheckpointCoordinator.java:693)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:930)
>   ... 8 more
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35159) CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash

2024-04-19 Thread Chesnay Schepler (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838908#comment-17838908
 ] 

Chesnay Schepler edited comment on FLINK-35159 at 4/19/24 3:03 PM:
---

master: 131358b918ba511064c23dfee96969a628d07d2a
1.19: 87ed9ccc2103457ba91f6ca45adfd2bfcc75c9ac
1.18: aacc735806acf1d63fa732706e079bc2ca1bb4fc


was (Author: zentol):
master: 131358b918ba511064c23dfee96969a628d07d2a
1.19:
1.18: aacc735806acf1d63fa732706e079bc2ca1bb4fc

> CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash
> 
>
> Key: FLINK-35159
> URL: https://issues.apache.org/jira/browse/FLINK-35159
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.18.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.2, 1.20.0, 1.19.1
>
>
> When a task manager dies while the JM is generating an ExecutionGraph in the 
> background then {{CreatingExecutionGraph#handleExecutionGraphCreation}} can 
> transition back into WaitingForResources if the TM hosted one of the slots 
> that we planned to use in {{tryToAssignSlots}}.
> At this point the ExecutionGraph was already transitioned to running, which 
> implicitly kicks of periodic checkpointing by the CheckpointCoordinator, 
> without the operator coordinator holders being initialized yet (as this 
> happens after we assigned slots).
> This effectively leaks that CheckpointCoordinator, including the timer thread 
> that will continue to try triggering checkpoints, which will naturally fail 
> to trigger.
> This can cause a JM crash because it results in 
> {{OperatorCoordinatorHolder#abortCurrentTriggering}} to be called, which 
> fails with an NPE since the {{mainThreadExecutor}} was not initialized yet.
> {code}
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$8(CheckpointCoordinator.java:707)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:610)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:910)
>   at 
> java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478)
>   at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>   at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>   at 
> java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>   at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.util.concurrent.CompletionException: 
> java.lang.NullPointerException
>   at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
>   at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
>   ... 7 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.abortCurrentTriggering(OperatorCoordinatorHolder.java:388)
>   at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
>   at 
> java.base/java.util.Collections$UnmodifiableCollection.forEach(Collections.java:1085)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:985)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:961)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$7(CheckpointCoordinator.java:693)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:930)
>   ... 8 more
> {code}



--
This 

Re: [PR] [BP-1.19][FLINK-35159] Transition ExecutionGraph to RUNNING after slot assignment [flink]

2024-04-19 Thread via GitHub


zentol merged PR #24686:
URL: https://github.com/apache/flink/pull/24686


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-28048][connectors] Introduce Source API alternative to FiniteTestSource [flink]

2024-04-19 Thread via GitHub


afedulov commented on PR #23777:
URL: https://github.com/apache/flink/pull/23777#issuecomment-2066761675

   CI is green: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=58998=results


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33460][Connector/JDBC] Support property authentication connection. [flink-connector-jdbc]

2024-04-19 Thread via GitHub


RocMarshal commented on code in PR #115:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/115#discussion_r1572398050


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcConnectionOptions.java:
##
@@ -90,13 +98,31 @@ public JdbcConnectionOptionsBuilder withDriverName(String 
driverName) {
 return this;
 }
 
+public JdbcConnectionOptionsBuilder withProperties(Properties 
properties) {

Review Comment:
   Thanks @eskabetxe for the comments.
   That sounds nice to me 
   updated~



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33460][Connector/JDBC] Support property authentication connection. [flink-connector-jdbc]

2024-04-19 Thread via GitHub


RocMarshal commented on code in PR #115:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/115#discussion_r1572398050


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcConnectionOptions.java:
##
@@ -90,13 +98,31 @@ public JdbcConnectionOptionsBuilder withDriverName(String 
driverName) {
 return this;
 }
 
+public JdbcConnectionOptionsBuilder withProperties(Properties 
properties) {

Review Comment:
   Thanks for the comments.
   That sounds nice to me 
   updated~



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-35159) CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash

2024-04-19 Thread Chesnay Schepler (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35159?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838908#comment-17838908
 ] 

Chesnay Schepler edited comment on FLINK-35159 at 4/19/24 1:40 PM:
---

master: 131358b918ba511064c23dfee96969a628d07d2a
1.19:
1.18: aacc735806acf1d63fa732706e079bc2ca1bb4fc


was (Author: zentol):
master: 131358b918ba511064c23dfee96969a628d07d2a
1.19:
1.18:

> CreatingExecutionGraph can leak CheckpointCoordinator and cause JM crash
> 
>
> Key: FLINK-35159
> URL: https://issues.apache.org/jira/browse/FLINK-35159
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.18.0
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.18.2, 1.20.0, 1.19.1
>
>
> When a task manager dies while the JM is generating an ExecutionGraph in the 
> background then {{CreatingExecutionGraph#handleExecutionGraphCreation}} can 
> transition back into WaitingForResources if the TM hosted one of the slots 
> that we planned to use in {{tryToAssignSlots}}.
> At this point the ExecutionGraph was already transitioned to running, which 
> implicitly kicks of periodic checkpointing by the CheckpointCoordinator, 
> without the operator coordinator holders being initialized yet (as this 
> happens after we assigned slots).
> This effectively leaks that CheckpointCoordinator, including the timer thread 
> that will continue to try triggering checkpoints, which will naturally fail 
> to trigger.
> This can cause a JM crash because it results in 
> {{OperatorCoordinatorHolder#abortCurrentTriggering}} to be called, which 
> fails with an NPE since the {{mainThreadExecutor}} was not initialized yet.
> {code}
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$8(CheckpointCoordinator.java:707)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:986)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:970)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:610)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:910)
>   at 
> java.base/java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:478)
>   at 
> java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
>   at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
>   at 
> java.base/java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>   at 
> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>   at java.base/java.lang.Thread.run(Thread.java:829)
> Caused by: java.util.concurrent.CompletionException: 
> java.lang.NullPointerException
>   at 
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
>   at 
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:932)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:907)
>   ... 7 more
> Caused by: java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.abortCurrentTriggering(OperatorCoordinatorHolder.java:388)
>   at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
>   at 
> java.base/java.util.Collections$UnmodifiableCollection.forEach(Collections.java:1085)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:985)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.onTriggerFailure(CheckpointCoordinator.java:961)
>   at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.lambda$startTriggeringCheckpoint$7(CheckpointCoordinator.java:693)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:930)
>   ... 8 more
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [BP-1.19][FLINK-35159] Transition ExecutionGraph to RUNNING after slot assignment [flink]

2024-04-19 Thread via GitHub


zentol commented on PR #24686:
URL: https://github.com/apache/flink/pull/24686#issuecomment-2066614601

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [BP-1.18][FLINK-35159] Transition ExecutionGraph to RUNNING after slot assignment [flink]

2024-04-19 Thread via GitHub


zentol merged PR #24687:
URL: https://github.com/apache/flink/pull/24687


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35175) HadoopDataInputStream can't compile with Hadoop 3.2.3

2024-04-19 Thread Ryan Skraba (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17839001#comment-17839001
 ] 

Ryan Skraba commented on FLINK-35175:
-

1.20 Hadoop 3.1.3 / Compile 
https://github.com/apache/flink/actions/runs/8747381080/job/24005737445#step:6:1560

> HadoopDataInputStream can't compile with Hadoop 3.2.3
> -
>
> Key: FLINK-35175
> URL: https://issues.apache.org/jira/browse/FLINK-35175
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.20.0
>Reporter: Ryan Skraba
>Priority: Critical
>
> Unfortunately, introduced in FLINK-35045: 
> [PREADWRITEBUFFER|https://github.com/apache/flink/commit/a312a3bdd258e0ff7d6f94e979b32e2bc762b82f#diff-3ed57be01895ba0f792110e40f4283427c55528f11a5105b4bf34ebd4e6fef0dR182]
>  was added in Hadoop releases 
> [3.3.0|https://github.com/apache/hadoop/blob/rel/release-3.3.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72]
>  and 
> [2.10.0|https://github.com/apache/hadoop/blob/rel/release-2.10.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72].
> It doesn't exist in flink.hadoop.version 
> [3.2.3|https://github.com/apache/hadoop/blob/rel/release-3.2.3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java],
>  which we are using in end-to-end tests.
> {code:java}
> 00:23:55.093 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.8.0:compile 
> (default-compile) on project flink-hadoop-fs: Compilation failure: 
> Compilation failure: 
> 00:23:55.093 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[151,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable READBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[182,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable PREADBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[183,43]
>  incompatible types: long cannot be converted to 
> org.apache.hadoop.io.ByteBufferPool
> 00:23:55.094 [ERROR] -> [Help 1] {code}
> * 1.20 compile_cron_hadoop313 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34227) Job doesn't disconnect from ResourceManager

2024-04-19 Thread Ryan Skraba (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34227?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17839002#comment-17839002
 ] 

Ryan Skraba commented on FLINK-34227:
-

1.18 AdaptiveScheduler / Test (module: table) 
https://github.com/apache/flink/actions/runs/8747381390/job/24006045095#step:10:12211

> Job doesn't disconnect from ResourceManager
> ---
>
> Key: FLINK-34227
> URL: https://issues.apache.org/jira/browse/FLINK-34227
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.19.0, 1.18.1
>Reporter: Matthias Pohl
>Assignee: Matthias Pohl
>Priority: Critical
>  Labels: github-actions, pull-request-available, test-stability
> Attachments: FLINK-34227.7e7d69daebb438b8d03b7392c9c55115.log, 
> FLINK-34227.log
>
>
> https://github.com/XComp/flink/actions/runs/7634987973/job/20800205972#step:10:14557
> {code}
> [...]
> "main" #1 prio=5 os_prio=0 tid=0x7f4b7000 nid=0x24ec0 waiting on 
> condition [0x7fccce1eb000]
>java.lang.Thread.State: WAITING (parking)
>   at sun.misc.Unsafe.park(Native Method)
>   - parking to wait for  <0xbdd52618> (a 
> java.util.concurrent.CompletableFuture$Signaller)
>   at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
>   at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
>   at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
>   at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
>   at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:2131)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:2099)
>   at 
> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.java:2077)
>   at 
> org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.execute(StreamExecutionEnvironment.scala:876)
>   at 
> org.apache.flink.table.planner.runtime.stream.sql.WindowDistinctAggregateITCase.testHopWindow_Cube(WindowDistinctAggregateITCase.scala:550)
> [...]
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-18476) PythonEnvUtilsTest#testStartPythonProcess fails

2024-04-19 Thread Ryan Skraba (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18476?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17839003#comment-17839003
 ] 

Ryan Skraba commented on FLINK-18476:
-

1.19 Java 11 / Test (module: misc) 
https://github.com/apache/flink/actions/runs/8747381464/job/24006015161#step:10:21789

> PythonEnvUtilsTest#testStartPythonProcess fails
> ---
>
> Key: FLINK-18476
> URL: https://issues.apache.org/jira/browse/FLINK-18476
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Tests
>Affects Versions: 1.11.0, 1.15.3, 1.18.0, 1.19.0, 1.20.0
>Reporter: Dawid Wysakowicz
>Priority: Major
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> test-stability
>
> The 
> {{org.apache.flink.client.python.PythonEnvUtilsTest#testStartPythonProcess}} 
> failed in my local environment as it assumes the environment has 
> {{/usr/bin/python}}. 
> I don't know exactly how did I get python in Ubuntu 20.04, but I have only 
> alias for {{python = python3}}. Therefore the tests fails.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-35175) HadoopDataInputStream can't compile with Hadoop 3.2.3

2024-04-19 Thread Ryan Skraba (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838999#comment-17838999
 ] 

Ryan Skraba commented on FLINK-35175:
-

[~hangxiang]: what do you think? Is it better to bump the supported version of 
Hadoop or fix the class that relies on this capability?

> HadoopDataInputStream can't compile with Hadoop 3.2.3
> -
>
> Key: FLINK-35175
> URL: https://issues.apache.org/jira/browse/FLINK-35175
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.20.0
>Reporter: Ryan Skraba
>Priority: Critical
>
> Unfortunately, introduced in FLINK-35045: 
> [PREADWRITEBUFFER|https://github.com/apache/flink/commit/a312a3bdd258e0ff7d6f94e979b32e2bc762b82f#diff-3ed57be01895ba0f792110e40f4283427c55528f11a5105b4bf34ebd4e6fef0dR182]
>  was added in Hadoop releases 
> [3.3.0|https://github.com/apache/hadoop/blob/rel/release-3.3.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72]
>  and 
> [2.10.0|https://github.com/apache/hadoop/blob/rel/release-2.10.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72].
> It doesn't exist in flink.hadoop.version 
> [3.2.3|https://github.com/apache/hadoop/blob/rel/release-3.2.3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java],
>  which we are using in end-to-end tests.
> {code:java}
> 00:23:55.093 [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-compiler-plugin:3.8.0:compile 
> (default-compile) on project flink-hadoop-fs: Compilation failure: 
> Compilation failure: 
> 00:23:55.093 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[151,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable READBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[182,63]
>  cannot find symbol
> 00:23:55.094 [ERROR]   symbol:   variable PREADBYTEBUFFER
> 00:23:55.094 [ERROR]   location: interface 
> org.apache.hadoop.fs.StreamCapabilities
> 00:23:55.094 [ERROR] 
> /home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[183,43]
>  incompatible types: long cannot be converted to 
> org.apache.hadoop.io.ByteBufferPool
> 00:23:55.094 [ERROR] -> [Help 1] {code}
> * 1.20 compile_cron_hadoop313 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35045][state] Introduce ForStFlinkFileSystem to support reading and writing with ByteBuffer [flink]

2024-04-19 Thread via GitHub


RyanSkraba commented on code in PR #24632:
URL: https://github.com/apache/flink/pull/24632#discussion_r1572359539


##
flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:
##
@@ -140,4 +144,56 @@ public void skipFully(long bytes) throws IOException {
 bytes -= fsDataInputStream.skip(bytes);
 }
 }
+
+@Override
+public int read(ByteBuffer byteBuffer) throws IOException {
+// Not all internal stream supports ByteBufferReadable
+if 
(fsDataInputStream.hasCapability(StreamCapabilities.READBYTEBUFFER)) {
+return fsDataInputStream.read(byteBuffer);
+} else {
+if (byteBuffer.hasArray()) {
+int len = byteBuffer.remaining();
+fsDataInputStream.readFully(byteBuffer.array(), 
byteBuffer.arrayOffset(), len);
+return len;
+} else {
+// Fallback to read byte then put
+int c = read();
+if (c == -1) {
+return -1;
+}
+byteBuffer.put((byte) c);
+
+int n = 1, len = byteBuffer.remaining() + 1;
+for (; n < len; n++) {
+c = read();
+if (c == -1) {
+break;
+}
+byteBuffer.put((byte) c);
+}
+return n;
+}
+}
+}
+
+@Override
+public int read(long position, ByteBuffer byteBuffer) throws IOException {
+// Not all internal stream supports ByteBufferPositionedReadable
+if 
(fsDataInputStream.hasCapability(StreamCapabilities.PREADBYTEBUFFER)) {

Review Comment:
   Hello!  Unfortunately this causes a compile error when using Hadoop 3.2.3 
(found in the nightly builds).  I created  FLINK-35175 to address this.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35175) HadoopDataInputStream can't compile with Hadoop 3.2.3

2024-04-19 Thread Ryan Skraba (Jira)
Ryan Skraba created FLINK-35175:
---

 Summary: HadoopDataInputStream can't compile with Hadoop 3.2.3
 Key: FLINK-35175
 URL: https://issues.apache.org/jira/browse/FLINK-35175
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.20.0
Reporter: Ryan Skraba


Unfortunately, introduced in FLINK-35045: 
[PREADWRITEBUFFER|https://github.com/apache/flink/commit/a312a3bdd258e0ff7d6f94e979b32e2bc762b82f#diff-3ed57be01895ba0f792110e40f4283427c55528f11a5105b4bf34ebd4e6fef0dR182]
 was added in Hadoop releases 
[3.3.0|https://github.com/apache/hadoop/blob/rel/release-3.3.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72]
 and 
[2.10.0|https://github.com/apache/hadoop/blob/rel/release-2.10.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java#L72].

It doesn't exist in flink.hadoop.version 
[3.2.3|https://github.com/apache/hadoop/blob/rel/release-3.2.3/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/StreamCapabilities.java],
 which we are using in end-to-end tests.
{code:java}
00:23:55.093 [ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-compiler-plugin:3.8.0:compile (default-compile) 
on project flink-hadoop-fs: Compilation failure: Compilation failure: 
00:23:55.093 [ERROR] 
/home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[151,63]
 cannot find symbol
00:23:55.094 [ERROR]   symbol:   variable READBYTEBUFFER
00:23:55.094 [ERROR]   location: interface 
org.apache.hadoop.fs.StreamCapabilities
00:23:55.094 [ERROR] 
/home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[182,63]
 cannot find symbol
00:23:55.094 [ERROR]   symbol:   variable PREADBYTEBUFFER
00:23:55.094 [ERROR]   location: interface 
org.apache.hadoop.fs.StreamCapabilities
00:23:55.094 [ERROR] 
/home/vsts/work/1/s/flink-filesystems/flink-hadoop-fs/src/main/java/org/apache/flink/runtime/fs/hdfs/HadoopDataInputStream.java:[183,43]
 incompatible types: long cannot be converted to 
org.apache.hadoop.io.ByteBufferPool
00:23:55.094 [ERROR] -> [Help 1] {code}
* 1.20 compile_cron_hadoop313 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59012=logs=87489130-75dc-54e4-1f45-80c30aa367a3=73da6d75-f30d-5d5a-acbe-487a9dcff678=3630



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35026][runtime][config] Introduce async execution configurations [flink]

2024-04-19 Thread via GitHub


yunfengzhou-hub commented on code in PR #24667:
URL: https://github.com/apache/flink/pull/24667#discussion_r1572344539


##
flink-core/src/main/java/org/apache/flink/configuration/ExecutionOptions.java:
##
@@ -181,4 +182,62 @@ public class ExecutionOptions {
 + " operators. NOTE: It takes effect only 
in the BATCH runtime mode and requires sorted inputs"
 + SORT_INPUTS.key()
 + " to be enabled.");
+
+// - Async State Execution 
--
+
+/**
+ * The max limit of in-flight records number in async state execution, 
'in-flight' refers to the
+ * records that have entered the operator but have not yet been processed 
and emitted to the
+ * downstream. If the in-flight records number exceeds the limit, the 
newly records entering
+ * will be blocked until the in-flight records number drops below the 
limit.
+ */
+@Experimental
+@Documentation.ExcludeFromDocumentation(
+"This is an experimental option, internal use only for now.")
+public static final ConfigOption ASYNC_INFLIGHT_RECORDS_LIMIT =
+ConfigOptions.key("execution.async-state.in-flight-records-limit")
+.intType()
+.defaultValue(6000)
+.withDescription(
+"The max limit of in-flight records number in 
async state execution, 'in-flight' refers"
++ " to the records that have entered the 
operator but have not yet been processed and"
++ " emitted to the downstream. If the 
in-flight records number exceeds the limit,"
++ " the newly records entering will be 
blocked until the in-flight records number drops below the limit.");
+
+/**
+ * The size of buffer under async state execution. Async state execution 
provides a buffer
+ * mechanism to reduce state access. When the number of state requests in 
the buffer exceeds the
+ * batch size, a batched state execution would be triggered. Larger batch 
sizes will bring
+ * higher end-to-end latency, this option works with {@link 
#ASYNC_STATE_BUFFER_TIMEOUT} to
+ * control the frequency of triggering.
+ */
+@Experimental
+@Documentation.ExcludeFromDocumentation(
+"This is an experimental option, internal use only for now.")
+public static final ConfigOption ASYNC_STATE_BUFFER_SIZE =
+ConfigOptions.key("execution.async-state.buffer-size")
+.intType()
+.defaultValue(1000)
+.withDescription(
+"The size of buffer under async state execution. 
Async state execution provides a buffer mechanism to reduce state access."
++ " When the number of state requests in 
the active buffer exceeds the batch size,"
++ " a batched state execution would be 
triggered. Larger batch sizes will bring higher end-to-end latency,"
++ " this option works with 
'execution.async-state.buffer-timeout' to control the frequency of 
triggering.");
+
+/**
+ * The timeout of buffer triggering in milliseconds. If the buffer has not 
reached the {@link
+ * #ASYNC_STATE_BUFFER_SIZE} within 'buffer-timeout' milliseconds, a 
trigger will perform
+ * actively.
+ */
+@Experimental
+@Documentation.ExcludeFromDocumentation(
+"This is an experimental option, internal use only for now.")
+public static final ConfigOption ASYNC_STATE_BUFFER_TIMEOUT =

Review Comment:
   It might be better to make time configurations as `long` instead of as `int`.



##
flink-runtime/src/main/java/org/apache/flink/runtime/asyncprocessing/AsyncExecutionController.java:
##
@@ -50,15 +50,24 @@ public class AsyncExecutionController {
 
 private static final Logger LOG = 
LoggerFactory.getLogger(AsyncExecutionController.class);
 
-public static final int DEFAULT_BATCH_SIZE = 1000;
-public static final int DEFAULT_MAX_IN_FLIGHT_RECORD_NUM = 6000;
+private static final int DEFAULT_BATCH_SIZE = 1000;
+
+private static final int DEFAULT_BUFFER_TIMEOUT = 1000;
+private static final int DEFAULT_MAX_IN_FLIGHT_RECORD_NUM = 6000;
 
 /**
  * The batch size. When the number of state requests in the active buffer 
exceeds the batch
  * size, a batched state execution would be triggered.
  */
 private final int batchSize;
 
+/**
+ * The timeout of {@link StateRequestBuffer#activeQueue} triggering in 
milliseconds. If the
+ * activeQueue has not reached the {@link #batchSize} within 
'buffer-timeout' milliseconds, a
+ * trigger will perform actively.
+ */
+private final int bufferTimeOut;


Re: [PR] [FLINK-35028][runtime] Timer firing under async execution model [flink]

2024-04-19 Thread via GitHub


yunfengzhou-hub commented on code in PR #24672:
URL: https://github.com/apache/flink/pull/24672#discussion_r1572300164


##
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/operators/asyncprocessing/AbstractAsyncStateStreamOperatorV2.java:
##
@@ -113,6 +120,26 @@ public final  ThrowingConsumer, 
Exception> getRecordProcessor
 + " since this part is handled by the Input.");
 }
 
+@Override
+@SuppressWarnings("unchecked")
+public  InternalTimerService getInternalTimerService(
+String name, TypeSerializer namespaceSerializer, Triggerable triggerable) {

Review Comment:
   A fallback plan when isAsyncStateProcessingEnabled=false might be needed 
here as well.



##
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/InternalTimerServiceAsyncImpl.java:
##
@@ -0,0 +1,146 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.operators;
+
+import org.apache.flink.runtime.asyncprocessing.AsyncExecutionController;
+import org.apache.flink.runtime.asyncprocessing.RecordContext;
+import org.apache.flink.runtime.metrics.groups.TaskIOMetricGroup;
+import org.apache.flink.runtime.state.KeyGroupRange;
+import org.apache.flink.runtime.state.KeyGroupedInternalPriorityQueue;
+import org.apache.flink.streaming.runtime.tasks.ProcessingTimeService;
+import org.apache.flink.streaming.runtime.tasks.StreamTaskCancellationContext;
+import org.apache.flink.util.CloseableIterator;
+import org.apache.flink.util.function.BiConsumerWithException;
+
+import static 
org.apache.flink.runtime.asyncprocessing.KeyAccountingUnit.EMPTY_RECORD;
+
+/**
+ * An implementation of {@link InternalTimerService} that is used by {@link
+ * 
org.apache.flink.streaming.runtime.operators.asyncprocessing.AbstractAsyncStateStreamOperator}.
+ * The timer service will set {@link RecordContext} for the timers before 
invoking action to
+ * preserve the execution order between timer firing and records processing.
+ *
+ * @see https://cwiki.apache.org/confluence/display/FLINK/FLIP-425%3A+Asynchronous+Execution+Model#FLIP425:AsynchronousExecutionModel-Timers>FLIP-425
+ * timers section.
+ * @param  Type of timer's key.
+ * @param  Type of the namespace to which timers are scoped.
+ */
+public class InternalTimerServiceAsyncImpl extends 
InternalTimerServiceImpl {
+
+private AsyncExecutionController asyncExecutionController;
+
+InternalTimerServiceAsyncImpl(
+TaskIOMetricGroup taskIOMetricGroup,
+KeyGroupRange localKeyGroupRange,
+KeyContext keyContext,
+ProcessingTimeService processingTimeService,
+KeyGroupedInternalPriorityQueue processingTimeTimersQueue,
+KeyGroupedInternalPriorityQueue eventTimeTimersQueue,
+StreamTaskCancellationContext cancellationContext,
+AsyncExecutionController asyncExecutionController) {
+super(
+taskIOMetricGroup,
+localKeyGroupRange,
+keyContext,
+processingTimeService,
+processingTimeTimersQueue,
+eventTimeTimersQueue,
+cancellationContext);
+this.asyncExecutionController = asyncExecutionController;
+this.processingTimeCallback = this::onProcessingTime;
+}
+
+private void onProcessingTime(long time) throws Exception {
+// null out the timer in case the Triggerable calls 
registerProcessingTimeTimer()
+// inside the callback.
+nextTimer = null;
+
+InternalTimer timer;
+
+while ((timer = processingTimeTimersQueue.peek()) != null
+&& timer.getTimestamp() <= time
+&& !cancellationContext.isCancelled()) {
+RecordContext recordCtx =
+asyncExecutionController.buildContext(EMPTY_RECORD, 
timer.getKey());
+recordCtx.retain();
+asyncExecutionController.setCurrentContext(recordCtx);
+keyContext.setCurrentKey(timer.getKey());
+processingTimeTimersQueue.poll();
+final InternalTimer timerToTrigger = timer;
+   

Re: [PR] [FLINK-34444] Initial implementation of JM operator metric rest api [flink]

2024-04-19 Thread via GitHub


afedulov commented on code in PR #24564:
URL: https://github.com/apache/flink/pull/24564#discussion_r1571313741


##
flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/legacy/metrics/MetricStore.java:
##
@@ -184,6 +184,21 @@ public synchronized ComponentMetricStore 
getJobManagerMetricStore() {
 return ComponentMetricStore.unmodifiable(jobManager);
 }
 
+public synchronized ComponentMetricStore getJobManagerOperatorMetricStore(
+String jobID, String taskID) {
+JobMetricStore job = jobID == null ? null : jobs.get(jobID);
+if (job == null || taskID == null) {
+return null;
+}
+
+TaskMetricStore task = job.getTaskMetricStore(taskID);
+if (task == null) {
+return null;
+}
+
+return 
ComponentMetricStore.unmodifiable(task.getJobManagerOperatorMetricStore());
+}

Review Comment:
   nit: feels like this could be a bit more readable
   ```suggestion
public synchronized ComponentMetricStore 
getJobManagerOperatorMetricStore(String jobID, String taskID) {
   if (jobID == null || taskID == null) {
   return null;
   }
   JobMetricStore job = jobs.get(jobID);
   if (job == null) {
   return null;
   }
   TaskMetricStore task = job.getTaskMetricStore(taskID);
   if (task == null) {
   return null;
   }
   return 
ComponentMetricStore.unmodifiable(task.getJobManagerOperatorMetricStore());
   }
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34444] Initial implementation of JM operator metric rest api [flink]

2024-04-19 Thread via GitHub


afedulov commented on code in PR #24564:
URL: https://github.com/apache/flink/pull/24564#discussion_r1572305374


##
docs/layouts/shortcodes/generated/rest_v1_dispatcher.html:
##
@@ -1277,6 +1277,9 @@
   "parallelism" : {
 "type" : "integer"
   },
+  "slotSharingGroupId" : {

Review Comment:
   Do I get it correctly that this is the consequence of [this 
commit](https://github.com/apache/flink/pull/24380/commits/6dec90667d24b4f197ad9a814f30b99ffe77a6af)
 not having regenerated the docs? If so, it could be better to have a separate 
[hotfix] commit in this PR for clarity.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-32151) 'Run kubernetes pyflink application test' fails while pulling image

2024-04-19 Thread Ryan Skraba (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838980#comment-17838980
 ] 

Ryan Skraba commented on FLINK-32151:
-

1.20 e2e_1_c1 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=59021=logs=bea52777-eaf8-5663-8482-18fbc3630e81=43ba8ce7-ebbf-57cd-9163-444305d74117=10878

> 'Run kubernetes pyflink application test' fails while pulling image
> ---
>
> Key: FLINK-32151
> URL: https://issues.apache.org/jira/browse/FLINK-32151
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Deployment / Kubernetes
>Affects Versions: 1.16.2
>Reporter: Sergey Nuyanzin
>Priority: Minor
>  Labels: auto-deprioritized-major, test-stability
>
> {noformat}
> 2023-05-16T13:29:39.0614891Z May 16 13:29:39 Current logs for 
> flink-native-k8s-pyflink-application-1-6f4c9bfc56-cstw7: 
> 2023-05-16T13:29:39.1253736Z Error from server (BadRequest): container 
> "flink-main-container" in pod 
> "flink-native-k8s-pyflink-application-1-6f4c9bfc56-cstw7" is waiting to 
> start: image can't be pulled
> 2023-05-16T13:29:39.2611218Z May 16 13:29:39 deployment.apps 
> "flink-native-k8s-pyflink-application-1" deleted
> 2023-05-16T13:29:39.4214711Z May 16 13:29:39 
> clusterrolebinding.rbac.authorization.k8s.io "flink-role-binding-default" 
> deleted
> 2023-05-16T13:29:40.2644587Z May 16 13:29:40 
> pod/flink-native-k8s-pyflink-application-1-6f4c9bfc56-cstw7 condition met
> 2023-05-16T13:29:40.2664618Z May 16 13:29:40 Stopping minikube ...
> 2023-05-16T13:29:40.3396336Z May 16 13:29:40 * Stopping node "minikube"  ...
> 2023-05-16T13:29:50.7499872Z May 16 13:29:50 * 1 node stopped.
> {noformat}
> it's very similar to https://issues.apache.org/jira/browse/FLINK-28226



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35133) Release flink-connector-cassandra v3.2.0 for Flink 1.19

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35133?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35133:
---
Labels: pull-request-available  (was: )

> Release flink-connector-cassandra v3.2.0 for Flink 1.19
> ---
>
> Key: FLINK-35133
> URL: https://issues.apache.org/jira/browse/FLINK-35133
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Cassandra
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
>
> https://github.com/apache/flink-connector-cassandra



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-35133][Connectors/Cassandra] Adding support for Flink 1.19 [flink-connector-cassandra]

2024-04-19 Thread via GitHub


dannycranmer opened a new pull request, #28:
URL: https://github.com/apache/flink-connector-cassandra/pull/28

   (no comment)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35069) ContinuousProcessingTimeTrigger continuously registers timers in a loop at the end of the window

2024-04-19 Thread Hong Liang Teoh (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838962#comment-17838962
 ] 

Hong Liang Teoh commented on FLINK-35069:
-

Thanks [~lijinzhong] for the fix! Are we planning to backport this fix to 1.18 
and 1.19 as well?

> ContinuousProcessingTimeTrigger continuously registers timers in a loop at 
> the end of the window
> 
>
> Key: FLINK-35069
> URL: https://issues.apache.org/jira/browse/FLINK-35069
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.17.0
>Reporter: Jinzhong Li
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-04-09-20-23-54-415.png
>
>
> In our production environment,  when TumblingEventTimeWindows and 
> ContinuousProcessingTimeTrigger are used in combination within a 
> WindowOperator, we observe a situation where the timers are continuously 
> registered in a loop at the end of the window, leading to the job being 
> perpetually stuck in timer processing.
> !image-2024-04-09-20-23-54-415.png|width=516,height=205!
> This issue can be reproduced using the 
> [UT|https://github.com/apache/flink/blob/8e80ff889701ed1abbb5c15cd3943b254f1fb5cc/flink-streaming-java/src/test/java/org/apache/flink/streaming/runtime/operators/windowing/ContinuousProcessingTimeTriggerTest.java#L177]
>  provided by the pr.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35168][State] Basic State Iterator for async processing [flink]

2024-04-19 Thread via GitHub


flinkbot commented on PR #24690:
URL: https://github.com/apache/flink/pull/24690#issuecomment-2066354887

   
   ## CI report:
   
   * 5fa3b48c1caed9d9e4ff9b31c1b8640318c18555 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35168) Basic State Iterator for async processing

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35168?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35168:
---
Labels: pull-request-available  (was: )

> Basic State Iterator for async processing
> -
>
> Key: FLINK-35168
> URL: https://issues.apache.org/jira/browse/FLINK-35168
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / State Backends
>Reporter: Zakelly Lan
>Assignee: Zakelly Lan
>Priority: Major
>  Labels: pull-request-available
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-35168][State] Basic State Iterator for async processing [flink]

2024-04-19 Thread via GitHub


Zakelly opened a new pull request, #24690:
URL: https://github.com/apache/flink/pull/24690

   ## What is the purpose of the change
   
   This PR ship the basic logic of state iterator, only leave the necessary 
part abstract for subclass to implement. Most of the code is future-related.
   
   ## Brief change log
   
- Introduce `AbstractStateIterator` and corresponding tests.
   
   
   ## Verifying this change
   
- Run newly added `AbstractStateIteratorTest`.
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): yes
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35174?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838954#comment-17838954
 ] 

Danny Cranmer commented on FLINK-35174:
---

Merged commit 66e323a into apache:main

> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
> ---
>
> Key: FLINK-35174
> URL: https://issues.apache.org/jira/browse/FLINK-35174
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: rabbitmq-3.1.0
>
>
> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35039][rest] Use PUT method supported by YARN web proxy instead of POST [flink]

2024-04-19 Thread via GitHub


Myasuka commented on PR #24689:
URL: https://github.com/apache/flink/pull/24689#issuecomment-2066319875

   @yeezychao Can your solution also works in standalone environment?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35174][Connectors/RabbitMQ] Bump commons-compress to 1.26.1 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer merged PR #28:
URL: https://github.com/apache/flink-connector-rabbitmq/pull/28


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35039) Create Profiling JobManager/TaskManager Instance failed

2024-04-19 Thread Yun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838951#comment-17838951
 ] 

Yun Tang commented on FLINK-35039:
--

[~wczhu] Already assigned to you.

> Create Profiling JobManager/TaskManager Instance failed
> ---
>
> Key: FLINK-35039
> URL: https://issues.apache.org/jira/browse/FLINK-35039
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend
>Affects Versions: 1.19.0
> Environment: Hadoop 3.2.2
> Flink 1.19
>Reporter: ude
>Assignee: ude
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-04-08-10-21-31-066.png, 
> image-2024-04-08-10-21-48-417.png, image-2024-04-08-10-30-16-683.png
>
>
> I'm test the "async-profiler" feature in version 1.19, but when I submit a 
> task in yarn per-job mode, I get an error  when I click Create Profiling 
> Instance on the flink Web UI page.
> !image-2024-04-08-10-21-31-066.png!
> !image-2024-04-08-10-21-48-417.png!
> The error message obviously means that the yarn proxy server does not support 
> *POST* calls. I checked the code of _*WebAppProxyServlet.java*_ and found 
> that the *POST* method is indeed not supported, so I changed it to *PUT* 
> method and the call was successful.
> !image-2024-04-08-10-30-16-683.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-35039) Create Profiling JobManager/TaskManager Instance failed

2024-04-19 Thread Yun Tang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yun Tang reassigned FLINK-35039:


Assignee: ude

> Create Profiling JobManager/TaskManager Instance failed
> ---
>
> Key: FLINK-35039
> URL: https://issues.apache.org/jira/browse/FLINK-35039
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend
>Affects Versions: 1.19.0
> Environment: Hadoop 3.2.2
> Flink 1.19
>Reporter: ude
>Assignee: ude
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-04-08-10-21-31-066.png, 
> image-2024-04-08-10-21-48-417.png, image-2024-04-08-10-30-16-683.png
>
>
> I'm test the "async-profiler" feature in version 1.19, but when I submit a 
> task in yarn per-job mode, I get an error  when I click Create Profiling 
> Instance on the flink Web UI page.
> !image-2024-04-08-10-21-31-066.png!
> !image-2024-04-08-10-21-48-417.png!
> The error message obviously means that the yarn proxy server does not support 
> *POST* calls. I checked the code of _*WebAppProxyServlet.java*_ and found 
> that the *POST* method is indeed not supported, so I changed it to *PUT* 
> method and the call was successful.
> !image-2024-04-08-10-30-16-683.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35055][cdc] Exclude connector dependencies with questionable licenses from fat jar packages [flink-cdc]

2024-04-19 Thread via GitHub


yuxiqian commented on PR #3212:
URL: https://github.com/apache/flink-cdc/pull/3212#issuecomment-2066306357

   Seems removing connector jar dependencies break SQL Job E2e tests. Need to 
fix this first.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35142][Connectors/RabbitMQ] Drop support for Flink 1.17 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer merged PR #27:
URL: https://github.com/apache/flink-connector-rabbitmq/pull/27


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [hotfix] Update copyright year to 2024 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer merged PR #24:
URL: https://github.com/apache/flink-connector-rabbitmq/pull/24


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35174][Connectors/RabbitMQ] Bump commons-compress to 1.26.1 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer commented on PR #28:
URL: 
https://github.com/apache/flink-connector-rabbitmq/pull/28#issuecomment-2066294353

   > Seems like there is a dependency convergence error
   > 
   > ```
   > Warning:  
   > Dependency convergence error for org.apache.commons:commons-lang3:3.14.0 
paths to dependency are:
   > +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
   >   +-org.testcontainers:junit-jupiter:1.19.3
   > +-org.testcontainers:testcontainers:1.19.3
   >   +-org.apache.commons:commons-compress:1.26.1
   > +-org.apache.commons:commons-lang3:3.14.0
   > and
   > +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
   >   
+-org.apache.flink:flink-architecture-tests-production:1.19-20240418.012917-199
   > +-org.apache.flink:flink-core:1.19-SNAPSHOT
   >   +-org.apache.commons:commons-lang3:3.12.0
   > and
   > +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
   >   
+-org.apache.flink:flink-architecture-tests-production:1.19-20240418.012917-199
   > +-org.apache.flink:flink-core:1.19-SNAPSHOT
   >   +-org.apache.commons:commons-text:1.10.0
   > +-org.apache.commons:commons-lang3:3.12.0
   > ```
   
   Thanks, fixed


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33460][Connector/JDBC] Support property authentication connection. [flink-connector-jdbc]

2024-04-19 Thread via GitHub


eskabetxe commented on code in PR #115:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/115#discussion_r1572162247


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcConnectionOptions.java:
##
@@ -90,13 +98,31 @@ public JdbcConnectionOptionsBuilder withDriverName(String 
driverName) {
 return this;
 }
 
+public JdbcConnectionOptionsBuilder withProperties(Properties 
properties) {

Review Comment:
   @RocMarshal I would say the easiest way, eliminating the method.
   
   If someone has a properties, they can do:
   `properties.forEach((key, value) -> builder.withProperty(key, value));`
   
   Under the hood, the `properties.putAll` is what it does.
   
   I always prefer to keep it simple..
   If we have to explain, document or add something to explain what is being 
done, we are doing it 'wrong'.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-35137) Release flink-connector-jdbc v3.2.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838591#comment-17838591
 ] 

Danny Cranmer edited comment on FLINK-35137 at 4/19/24 10:28 AM:
-

RC2 Vote: https://lists.apache.org/thread/b7xbjo4crt1527ldksw4nkwo8vs56csy


was (Author: dannycranmer):
https://lists.apache.org/thread/b7xbjo4crt1527ldksw4nkwo8vs56csy

> Release flink-connector-jdbc v3.2.0 for Flink 1.19
> --
>
> Key: FLINK-35137
> URL: https://issues.apache.org/jira/browse/FLINK-35137
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: jdbc-3.2.0
>
>
> https://github.com/apache/flink-connector-jdbc



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35139) Release flink-connector-mongodb v1.2.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838086#comment-17838086
 ] 

Danny Cranmer edited comment on FLINK-35139 at 4/19/24 10:28 AM:
-

RC1 Vote: https://lists.apache.org/thread/s18g7obgp4sbdtl73571976vqvy1ftk8


was (Author: dannycranmer):
https://lists.apache.org/thread/s18g7obgp4sbdtl73571976vqvy1ftk8

> Release flink-connector-mongodb v1.2.0 for Flink 1.19
> -
>
> Key: FLINK-35139
> URL: https://issues.apache.org/jira/browse/FLINK-35139
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / MongoDB
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: mongodb-1.2.0
>
>
> https://github.com/apache/flink-connector-mongodb



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35137) Release flink-connector-jdbc v3.2.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35137?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838181#comment-17838181
 ] 

Danny Cranmer edited comment on FLINK-35137 at 4/19/24 10:28 AM:
-

RC1 Vote: https://lists.apache.org/thread/q6dmc5dbz7kcfvpo99pj2sh5mzhffgl5


was (Author: dannycranmer):
https://lists.apache.org/thread/q6dmc5dbz7kcfvpo99pj2sh5mzhffgl5

> Release flink-connector-jdbc v3.2.0 for Flink 1.19
> --
>
> Key: FLINK-35137
> URL: https://issues.apache.org/jira/browse/FLINK-35137
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / JDBC
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: jdbc-3.2.0
>
>
> https://github.com/apache/flink-connector-jdbc



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35132) Release flink-connector-aws v4.3.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838058#comment-17838058
 ] 

Danny Cranmer edited comment on FLINK-35132 at 4/19/24 10:27 AM:
-

RC1 Vote: [https://lists.apache.org/thread/0nw9smt23crx4gwkf6p1dd4jwvp1g5s0]


was (Author: dannycranmer):
https://lists.apache.org/thread/0nw9smt23crx4gwkf6p1dd4jwvp1g5s0

> Release flink-connector-aws v4.3.0 for Flink 1.19
> -
>
> Key: FLINK-35132
> URL: https://issues.apache.org/jira/browse/FLINK-35132
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / AWS
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: aws-connector-4.3.0
>
>
> https://github.com/apache/flink-connector-aws



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35135) Release flink-connector-gcp-pubsub v3.1.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838590#comment-17838590
 ] 

Danny Cranmer edited comment on FLINK-35135 at 4/19/24 10:28 AM:
-

RC1 Vote: https://lists.apache.org/thread/b7l1r0y7nwox2vhf2z3kwjn41clf6w1v


was (Author: dannycranmer):
https://lists.apache.org/thread/b7l1r0y7nwox2vhf2z3kwjn41clf6w1v

> Release flink-connector-gcp-pubsub v3.1.0 for Flink 1.19
> 
>
> Key: FLINK-35135
> URL: https://issues.apache.org/jira/browse/FLINK-35135
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Google Cloud PubSub
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: gcp-pubsub-3.1.0
>
>
> https://github.com/apache/flink-connector-gcp-pubsub



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-35139) Release flink-connector-mongodb v1.2.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35139?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838592#comment-17838592
 ] 

Danny Cranmer edited comment on FLINK-35139 at 4/19/24 10:28 AM:
-

RC2 Vote: https://lists.apache.org/thread/2982v6n5q0bgldrp919t5t6d19xsl710


was (Author: dannycranmer):
https://lists.apache.org/thread/2982v6n5q0bgldrp919t5t6d19xsl710

> Release flink-connector-mongodb v1.2.0 for Flink 1.19
> -
>
> Key: FLINK-35139
> URL: https://issues.apache.org/jira/browse/FLINK-35139
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / MongoDB
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: mongodb-1.2.0
>
>
> https://github.com/apache/flink-connector-mongodb



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35174][Connectors/RabbitMQ] Bump commons-compress to 1.26.1 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


hlteoh37 commented on PR #28:
URL: 
https://github.com/apache/flink-connector-rabbitmq/pull/28#issuecomment-2066287030

   Seems like there is a dependency convergence error
   
   ```
   Warning:  
   Dependency convergence error for org.apache.commons:commons-lang3:3.14.0 
paths to dependency are:
   +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
 +-org.testcontainers:junit-jupiter:1.19.3
   +-org.testcontainers:testcontainers:1.19.3
 +-org.apache.commons:commons-compress:1.26.1
   +-org.apache.commons:commons-lang3:3.14.0
   and
   +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
 
+-org.apache.flink:flink-architecture-tests-production:1.19-20240418.012917-199
   +-org.apache.flink:flink-core:1.19-SNAPSHOT
 +-org.apache.commons:commons-lang3:3.12.0
   and
   +-org.apache.flink:flink-connector-rabbitmq-parent:3.1-SNAPSHOT
 
+-org.apache.flink:flink-architecture-tests-production:1.19-20240418.012917-199
   +-org.apache.flink:flink-core:1.19-SNAPSHOT
 +-org.apache.commons:commons-text:1.10.0
   +-org.apache.commons:commons-lang3:3.12.0
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35132) Release flink-connector-aws v4.3.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35132?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838949#comment-17838949
 ] 

Danny Cranmer commented on FLINK-35132:
---

RC2 Vote: https://lists.apache.org/thread/dof3tprw5s19jfwhk6yxs5qxbb4x5ss8

> Release flink-connector-aws v4.3.0 for Flink 1.19
> -
>
> Key: FLINK-35132
> URL: https://issues.apache.org/jira/browse/FLINK-35132
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / AWS
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: aws-connector-4.3.0
>
>
> https://github.com/apache/flink-connector-aws



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33460][Connector/JDBC] Support property authentication connection. [flink-connector-jdbc]

2024-04-19 Thread via GitHub


eskabetxe commented on code in PR #115:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/115#discussion_r1572162247


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcConnectionOptions.java:
##
@@ -90,13 +98,31 @@ public JdbcConnectionOptionsBuilder withDriverName(String 
driverName) {
 return this;
 }
 
+public JdbcConnectionOptionsBuilder withProperties(Properties 
properties) {

Review Comment:
   @RocMarshal I would say the easiest way, eliminating the method.
   
   If someone has a properties, they can do:
   `properties.forEach((key, value) -> builder.withProperty(key, value));`
   
   Under the hood, the `properties.putAll` is what it does.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35162][state] Implement WriteBatchOperation and general MultiGetOperation for ForSt [flink]

2024-04-19 Thread via GitHub


ljz2051 commented on PR #24681:
URL: https://github.com/apache/flink/pull/24681#issuecomment-2066281635

   @Zakelly Thanks for your review! I have addressed all comments your 
mentioned. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35162][state] Implement WriteBatchOperation and general MultiGetOperation for ForSt [flink]

2024-04-19 Thread via GitHub


ljz2051 commented on code in PR #24681:
URL: https://github.com/apache/flink/pull/24681#discussion_r1572160266


##
flink-state-backends/flink-statebackend-forst/src/main/java/org/apache/flink/state/forst/ForStDBOperation.java:
##
@@ -0,0 +1,42 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.state.forst;
+
+import org.apache.flink.annotation.Internal;
+
+import java.io.IOException;
+
+/**
+ * Data access operation to ForStDB. This interface is used to encapsulate the 
DB access operations
+ * formed after grouping state access. For more information about “Grouping 
state access”, please
+ * refer to FLIP-426.
+ *
+ * @param  The type of output for DB access.
+ */
+@Internal
+public interface ForStDBOperation {
+
+/**
+ * Process the ForStDB access requests.
+ *
+ * @return Processing result.
+ * @throws IOException Thrown if ForStDB access encountered an I/O related 
error.
+ */
+OUT process() throws IOException;

Review Comment:
   Good point. I agree that the future-style interface is more flexible. I have 
refined it.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35162][state] Implement WriteBatchOperation and general MultiGetOperation for ForSt [flink]

2024-04-19 Thread via GitHub


ljz2051 commented on code in PR #24681:
URL: https://github.com/apache/flink/pull/24681#discussion_r1572156946


##
flink-state-backends/flink-statebackend-forst/src/main/java/org/apache/flink/state/forst/ContextKey.java:
##
@@ -0,0 +1,73 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.state.forst;
+
+import javax.annotation.Nullable;
+
+/**
+ * The composite key which contains some context information.
+ *
+ * @param  The type of the raw key.
+ */
+public class ContextKey {

Review Comment:
   Yes,  the ContextKey can be shared across state requests, and ContextKey 
will be attached to the RecordContext in another PR. **So I move the ContextKey 
class to flink-runtime module from forst-module**.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35162][state] Implement WriteBatchOperation and general MultiGetOperation for ForSt [flink]

2024-04-19 Thread via GitHub


ljz2051 commented on code in PR #24681:
URL: https://github.com/apache/flink/pull/24681#discussion_r1572153463


##
flink-state-backends/flink-statebackend-forst/src/main/java/org/apache/flink/state/forst/ContextKey.java:
##
@@ -0,0 +1,73 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.state.forst;
+
+import javax.annotation.Nullable;
+
+/**
+ * The composite key which contains some context information.
+ *
+ * @param  The type of the raw key.
+ */
+public class ContextKey {
+
+private final K rawKey;
+
+private final int keyGroup;
+
+/**
+ * A record in user layer may access the state multiple times. The {@code 
serializedKey} can be
+ * used to cache the serialized key bytes after its first serialization, 
so that subsequent
+ * state accesses with the same key can avoid being serialized repeatedly.
+ */
+private @Nullable byte[] serializedKey = null;
+
+public ContextKey(K rawKey, int keyGroup) {

Review Comment:
   I have remove the "ContextKey#of" static method



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35142) Release flink-connector-rabbitmq v3.1.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35142:
--
Summary: Release flink-connector-rabbitmq v3.1.0 for Flink 1.19  (was: 
Release flink-connector-rabbitmq vX.X.X for Flink 1.19)

> Release flink-connector-rabbitmq v3.1.0 for Flink 1.19
> --
>
> Key: FLINK-35142
> URL: https://issues.apache.org/jira/browse/FLINK-35142
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
>
> https://github.com/apache/flink-connector-rabbitmq



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] Fix problem due to HBaseRowDataLookupFunction returned to reuse objects [flink-connector-hbase]

2024-04-19 Thread via GitHub


Tan-JiaLiang commented on PR #20:
URL: 
https://github.com/apache/flink-connector-hbase/pull/20#issuecomment-2066268025

   @Jackson-xky This is a serious problem in the hbase lookup join. Are you 
still working on it? I'd be happy to help.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.0 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dependabot[bot] commented on PR #25:
URL: 
https://github.com/apache/flink-connector-rabbitmq/pull/25#issuecomment-2066266561

   OK, I won't notify you again about this release, but will get in touch when 
a new version is available. If you'd rather skip all updates until the next 
major or minor version, let me know by commenting `@dependabot ignore this 
major version` or `@dependabot ignore this minor version`.
   
   If you change your mind, just re-open this PR and I'll resolve any conflicts 
on it.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.0 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer commented on PR #25:
URL: 
https://github.com/apache/flink-connector-rabbitmq/pull/25#issuecomment-2066266504

   https://issues.apache.org/jira/browse/FLINK-35174


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35162][state] Implement WriteBatchOperation and general MultiGetOperation for ForSt [flink]

2024-04-19 Thread via GitHub


ljz2051 commented on code in PR #24681:
URL: https://github.com/apache/flink/pull/24681#discussion_r1572151847


##
flink-state-backends/flink-statebackend-forst/src/main/java/org/apache/flink/state/forst/ForStValueState.java:
##
@@ -0,0 +1,99 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.state.forst;
+
+import org.apache.flink.api.common.state.v2.ValueState;
+import org.apache.flink.core.memory.DataInputDeserializer;
+import org.apache.flink.core.memory.DataOutputSerializer;
+import org.apache.flink.runtime.asyncprocessing.StateRequestHandler;
+import org.apache.flink.runtime.state.SerializedCompositeKeyBuilder;
+import org.apache.flink.runtime.state.v2.InternalValueState;
+import org.apache.flink.runtime.state.v2.ValueStateDescriptor;
+
+import org.rocksdb.ColumnFamilyHandle;
+
+import java.io.IOException;
+
+/**
+ * The {@link InternalValueState} implement for ForStDB.
+ *
+ * @param  The type of the key.
+ * @param  The type of the value.
+ */
+public class ForStValueState extends InternalValueState
+implements ValueState, ForStInnerTable, V> {
+
+/** The column family which this internal value state belongs to. */
+private final ColumnFamilyHandle columnFamilyHandle;
+
+/** The serialized key builder which need be thread-safe. */
+private final ThreadLocal> 
serializedKeyBuilder;
+
+/** The data outputStream used for value serializer, which need be 
thread-safe. */
+private final ThreadLocal valueSerializerView;
+
+/** The data inputStream used for value deserializer, which need be 
thread-safe. */
+private final ThreadLocal valueDeserializerView;
+
+public ForStValueState(
+StateRequestHandler stateRequestHandler,
+ColumnFamilyHandle columnFamily,
+ValueStateDescriptor valueStateDescriptor,
+ThreadLocal> serializedKeyBuilder,
+ThreadLocal valueSerializerView,
+ThreadLocal valueDeserializerView) {
+super(stateRequestHandler, valueStateDescriptor);
+this.columnFamilyHandle = columnFamily;
+this.serializedKeyBuilder = serializedKeyBuilder;
+this.valueSerializerView = valueSerializerView;
+this.valueDeserializerView = valueDeserializerView;
+}
+
+@Override
+public ColumnFamilyHandle getColumnFamilyHandle() {
+return columnFamilyHandle;
+}
+
+@Override
+public byte[] serializeKey(ContextKey contextKey) throws IOException {

Review Comment:
   Yes. I have marked the ContextKey with  @ThreadSafe annotation, and refactor 
the 'read cache or serialize' logic. Please review the 
ContextKey#getOrCreateSerializedKey method.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35174:
--
Description: 
Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
RabbitMQ connector

 

  was:Bump org.apache.commons:commons-compress from 1.25.0 to 1.26.1 for Flink 
AWS connectors


> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
> ---
>
> Key: FLINK-35174
> URL: https://issues.apache.org/jira/browse/FLINK-35174
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / AWS
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: aws-connector-4.3.0
>
>
> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35174:
--
Fix Version/s: rabbitmq-3.1.0
   (was: aws-connector-4.3.0)

> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
> ---
>
> Key: FLINK-35174
> URL: https://issues.apache.org/jira/browse/FLINK-35174
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / AWS
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: rabbitmq-3.1.0
>
>
> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35174:
--
Component/s: Connectors/ RabbitMQ
 (was: Connectors / AWS)

> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
> ---
>
> Key: FLINK-35174
> URL: https://issues.apache.org/jira/browse/FLINK-35174
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: rabbitmq-3.1.0
>
>
> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35174?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35174:
--
Summary: Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for 
Flink RabbitMQ connector  (was: Bump org.apache.commons:commons-compress from 
1.25.0 to 1.26.1 for Flink RabbitMQ connector)

> Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.1 for Flink 
> RabbitMQ connector
> ---
>
> Key: FLINK-35174
> URL: https://issues.apache.org/jira/browse/FLINK-35174
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / AWS
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: aws-connector-4.3.0
>
>
> Bump org.apache.commons:commons-compress from 1.25.0 to 1.26.1 for Flink AWS 
> connectors



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33460][Connector/JDBC] Support property authentication connection. [flink-connector-jdbc]

2024-04-19 Thread via GitHub


RocMarshal commented on code in PR #115:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/115#discussion_r1572083767


##
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcConnectionOptions.java:
##
@@ -90,13 +98,31 @@ public JdbcConnectionOptionsBuilder withDriverName(String 
driverName) {
 return this;
 }
 
+public JdbcConnectionOptionsBuilder withProperties(Properties 
properties) {

Review Comment:
   hi, @eskabetxe Thanks a lot for your quick-review  . 
   
   ```
   .withProperty("a", "a1")
   .withProperties(new properties())
   .withProperty("b", "b1")
   ``` 
   
   > .
   > that a and b should be in the properties, or just b...
   
   It's right.
   It's difficult to express , which depends the what cases devs need and I 
have to admit that the lines referenced above is ambiguous now.  
   
   I have come up with two temporary ways to alleviate this ambiguity:
   
   - Keep the current logic, and then we add comment to describe semantics for 
these two methods.
   
   OR
   
   - Add a flag parameter to method `withProperties(new properties(), 
true/false)` , indicating whether to force assign   entire `properties`, rather 
than a part of the `properties` by `upsert` mode.
   
   Of course, I'd appreciated to hear any ideas form you / community devs !  
   Please let me know what's your opinion.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35142) Release flink-connector-rabbitmq v3.1.0 for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-35142:
--
Fix Version/s: rabbitmq-3.1.0

> Release flink-connector-rabbitmq v3.1.0 for Flink 1.19
> --
>
> Key: FLINK-35142
> URL: https://issues.apache.org/jira/browse/FLINK-35142
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
> Fix For: rabbitmq-3.1.0
>
>
> https://github.com/apache/flink-connector-rabbitmq



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] Bump org.apache.commons:commons-compress from 1.24.0 to 1.26.0 [flink-connector-rabbitmq]

2024-04-19 Thread via GitHub


dannycranmer closed pull request #25: Bump org.apache.commons:commons-compress 
from 1.24.0 to 1.26.0
URL: https://github.com/apache/flink-connector-rabbitmq/pull/25


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-35174) Bump org.apache.commons:commons-compress from 1.25.0 to 1.26.1 for Flink RabbitMQ connector

2024-04-19 Thread Danny Cranmer (Jira)
Danny Cranmer created FLINK-35174:
-

 Summary: Bump org.apache.commons:commons-compress from 1.25.0 to 
1.26.1 for Flink RabbitMQ connector
 Key: FLINK-35174
 URL: https://issues.apache.org/jira/browse/FLINK-35174
 Project: Flink
  Issue Type: Technical Debt
  Components: Connectors / AWS
Reporter: Danny Cranmer
Assignee: Danny Cranmer
 Fix For: aws-connector-4.3.0


Bump org.apache.commons:commons-compress from 1.25.0 to 1.26.1 for Flink AWS 
connectors



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-35142) Release flink-connector-rabbitmq vX.X.X for Flink 1.19

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer reassigned FLINK-35142:
-

Assignee: Danny Cranmer

> Release flink-connector-rabbitmq vX.X.X for Flink 1.19
> --
>
> Key: FLINK-35142
> URL: https://issues.apache.org/jira/browse/FLINK-35142
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Assignee: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
>
> https://github.com/apache/flink-connector-rabbitmq



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-35142) Release flink-connector-rabbitmq vX.X.X for Flink 1.19

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35142:
---
Labels: pull-request-available  (was: )

> Release flink-connector-rabbitmq vX.X.X for Flink 1.19
> --
>
> Key: FLINK-35142
> URL: https://issues.apache.org/jira/browse/FLINK-35142
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors/ RabbitMQ
>Reporter: Danny Cranmer
>Priority: Major
>  Labels: pull-request-available
>
> https://github.com/apache/flink-connector-rabbitmq



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838937#comment-17838937
 ] 

Danny Cranmer commented on FLINK-34954:
---

Thanks for the bug report and fix, [~q.xu] .

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Assignee: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-34954:
--
Component/s: API / Type Serialization System

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer reassigned FLINK-34954:
-

Assignee: Qinghui Xu

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Assignee: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer resolved FLINK-34954.
---
Resolution: Fixed

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Assignee: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-34954:
--
Fix Version/s: 1.20.0

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>Reporter: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34954][core] Kryo Input bug fix [flink]

2024-04-19 Thread via GitHub


dannycranmer merged PR #24586:
URL: https://github.com/apache/flink/pull/24586


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-35039) Create Profiling JobManager/TaskManager Instance failed

2024-04-19 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35039?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-35039:
---
Labels: pull-request-available  (was: )

> Create Profiling JobManager/TaskManager Instance failed
> ---
>
> Key: FLINK-35039
> URL: https://issues.apache.org/jira/browse/FLINK-35039
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Web Frontend
>Affects Versions: 1.19.0
> Environment: Hadoop 3.2.2
> Flink 1.19
>Reporter: ude
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-04-08-10-21-31-066.png, 
> image-2024-04-08-10-21-48-417.png, image-2024-04-08-10-30-16-683.png
>
>
> I'm test the "async-profiler" feature in version 1.19, but when I submit a 
> task in yarn per-job mode, I get an error  when I click Create Profiling 
> Instance on the flink Web UI page.
> !image-2024-04-08-10-21-31-066.png!
> !image-2024-04-08-10-21-48-417.png!
> The error message obviously means that the yarn proxy server does not support 
> *POST* calls. I checked the code of _*WebAppProxyServlet.java*_ and found 
> that the *POST* method is indeed not supported, so I changed it to *PUT* 
> method and the call was successful.
> !image-2024-04-08-10-30-16-683.png!
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35039][rest] Use PUT method supported by YARN web proxy instead of POST [flink]

2024-04-19 Thread via GitHub


flinkbot commented on PR #24689:
URL: https://github.com/apache/flink/pull/24689#issuecomment-2066235806

   
   ## CI report:
   
   * 5179e513e55a60de15d50b32497442d432cf051b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838938#comment-17838938
 ] 

Danny Cranmer commented on FLINK-34954:
---

Merged commit 
[{{3977982}}|https://github.com/apache/flink/commit/39779829b8853c61c9da9eaf193f0fedf5857f7b]
 into apache:master

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>  Components: API / Type Serialization System
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Assignee: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34954) Kryo input implementation NoFetchingInput fails to handle zero length bytes

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-34954:
--
Affects Version/s: 1.19.0

> Kryo input implementation NoFetchingInput fails to handle zero length bytes
> ---
>
> Key: FLINK-34954
> URL: https://issues.apache.org/jira/browse/FLINK-34954
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.19.0
>Reporter: Qinghui Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>
> If the serailized bytes are empty, `NoFetchingInput` will run into error when 
> Kryo tries to deserialize it.
> Example: a protobuf 3 object that contains only default values will be 
> serialized as 0 length byte array, and the deserialization later will fail. 
> Illustration:
> {noformat}
> import com.esotericsoftware.kryo.Kryo
> import com.esotericsoftware.kryo.io.{ByteBufferInput, ByteBufferOutput, 
> Input, Output}
> import com.google.protobuf.{DescriptorProtos, Message}import 
> com.twitter.chill.protobuf.ProtobufSerializer
> import org.apache.flink.api.java.typeutils.runtime.NoFetchingInput
> import java.io.ByteArrayInputStream
>  
> object ProtoSerializationTest {
>   def main(args: Array[String]) = {     
> val chillProtoSerializer = new ProtobufSerializer
>     val protomessage = DescriptorProtos.DescriptorProto.getDefaultInstance
>     val output: Output = new ByteBufferOutput(1000)
>     chillProtoSerializer.write(null, output, protomessage)
>     val serialized: Array[Byte] = output.toBytes
>     println(s"Serialized : $serialized")
>     val input: Input = new NoFetchingInput(new 
> ByteArrayInputStream(serialized))
>     val deserialized = chillProtoSerializer.read(null, input, 
> classOf[BillableClick].asInstanceOf[Class[Message]])
>     println(deserialized)
>   }
> }
> {noformat}
>  
> Error
> {noformat}
> Exception in thread "main" java.lang.RuntimeException: Could not create class 
> com.criteo.glup.BillableClickProto$BillableClick
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:76)
>     at 
> com.criteo.streaming.common.bootstrap.ProtoSerialization$.main(ProtoSerialization.scala:22)
>     at ProtoSerialization.main(ProtoSerialization.scala)
> Caused by: com.esotericsoftware.kryo.KryoException: java.io.EOFException: No 
> more bytes left.
>     at 
> org.apache.flink.api.java.typeutils.runtime.NoFetchingInput.readBytes(NoFetchingInput.java:128)
>     at com.esotericsoftware.kryo.io.Input.readBytes(Input.java:332)
>     at 
> com.twitter.chill.protobuf.ProtobufSerializer.read(ProtobufSerializer.java:73)
>     ... 2 more
> Caused by: java.io.EOFException: No more bytes left.
>     ... 5 more{noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] Flink 35039 [flink]

2024-04-19 Thread via GitHub


yeezychao opened a new pull request, #24689:
URL: https://github.com/apache/flink/pull/24689

   
   
   ## What is the purpose of the change
   
   *(For example: This pull request makes task deployment go through the blob 
server, rather than through RPC. That way we avoid re-transferring them on each 
deployment (during recovery).)*
   
   
   ## Brief change log
   
   *(for example:)*
 - *The TaskInfo is stored in the blob store on job creation time as a 
persistent artifact*
 - *Deployments RPC transmits only the blob storage reference*
 - *TaskManagers retrieve the TaskInfo from the blob cache*
   
   
   ## Verifying this change
   
   Please make sure both new and modified tests in this PR follow [the 
conventions for tests defined in our code quality 
guide](https://flink.apache.org/how-to-contribute/code-style-and-quality-common/#7-testing).
   
   *(Please pick either of the following options)*
   
   This change is a trivial rework / code cleanup without any test coverage.
   
   *(or)*
   
   This change is already covered by existing tests, such as *(please describe 
tests)*.
   
   *(or)*
   
   This change added tests and can be verified as follows:
   
   *(example:)*
 - *Added integration tests for end-to-end deployment with large payloads 
(100MB)*
 - *Extended integration test for recovery after master (JobManager) 
failure*
 - *Added test that validates that TaskInfo is transferred only once across 
recoveries*
 - *Manually verified the change by running a 4 node cluster with 2 
JobManagers and 4 TaskManagers, a stateful streaming program, and killing one 
JobManager and two TaskManagers during the execution, verifying that recovery 
happens correctly.*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / no / 
don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
 - The S3 file system connector: (yes / no / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-35115) Kinesis connector writes wrong Kinesis sequence number at stop with savepoint

2024-04-19 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-35115?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer resolved FLINK-35115.
---
Resolution: Fixed

> Kinesis connector writes wrong Kinesis sequence number at stop with savepoint
> -
>
> Key: FLINK-35115
> URL: https://issues.apache.org/jira/browse/FLINK-35115
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.15.4, aws-connector-4.0.0, aws-connector-4.1.0, 
> aws-connector-4.2.0, 1.16.3, 1.17.2, 1.18.1
> Environment: The issue happens in a *Kinesis -> Flink -> Kafka* 
> exactly-once setup with:
>  * Flink versions checked 1.16.3 and 1.18.1
>  * Kinesis connector checked 1.16.3 and 4.2.0-1.18
>  * checkpointing configured at 1 minute with EXACTLY_ONCE mode: 
> {code:java}
> StreamExecutionEnvironment execEnv = 
> StreamExecutionEnvironment.getExecutionEnvironment (); 
> execEnv.enableCheckpointing (6,EXACTLY_ONCE); execEnv.getCheckpointConfig 
> ().setCheckpointTimeout (9); execEnv.getCheckpointConfig 
> ().setCheckpointStorage (CHECKPOINTS_PATH); {code}
>  * Kafka sink configured with EXACTLY_ONCE semantic/delivery guarantee:
> {code:java}
> Properties sinkConfig = new Properties ();
> sinkConfig.put ("transaction.timeout.ms", 48);
> KafkaSink sink = KafkaSink.builder ()
> .setBootstrapServers ("localhost:9092")
> .setTransactionalIdPrefix ("test-prefix")
> .setDeliverGuarantee (EXACTLY_ONCE)
> .setKafkaProducerConfig (sinkConfig)
> .setRecordSerializer (
> (KafkaRecordSerializationSchema) (element, context, 
> timestamp) -> new ProducerRecord<> (
> "test-output-topic", null, element.getBytes ()))
> .build (); {code}
>  * Kinesis consumer defined as: 
> {code:java}
> FlinkKinesisConsumer flinkKinesisConsumer = new
> FlinkKinesisConsumer<> ("test-stream",
> new AbstractDeserializationSchema<> () {
> @Override
> public ByteBuffer deserialize (byte[] bytes) {
> // Return
> return ByteBuffer.wrap (bytes);
> }
> }, props); {code}
>  
>Reporter: Vadim Vararu
>Assignee: Aleksandr Pilipenko
>Priority: Blocker
>  Labels: kinesis, pull-request-available
> Fix For: aws-connector-4.3.0
>
>
> Having an exactly-once Kinesis -> Flink -> Kafka job and triggering a 
> stop-with-savepoint, Flink duplicates in Kafka all the records between the 
> last checkpoint and the savepoint at resume:
>  * Event1 is written to Kinesis
>  * Event1 is processed by Flink 
>  * Event1 is committed to Kafka at the checkpoint
>  * 
> 
>  * Event2 is written to Kinesis
>  * Event2 is processed by Flink
>  * Stop with savepoint is triggered manually
>  * Event2 is committed to Kafka
>  * 
> 
>  * Job is resumed from the savepoint
>  * *{color:#FF}Event2 is written again to Kafka at the first 
> checkpoint{color}*
>  
> {color:#172b4d}I believe that it's a Kinesis connector issue for 2 
> reasons:{color}
>  * I've checked the actual Kinesis sequence number in the _metadata file 
> generated at stop-with-savepoint and it's the one from the checkpoint before 
> the savepoint  instead of being the one of the last record committed to Kafka.
>  * I've tested exactly the save job with Kafka as source instead of Kinesis 
> as source and the behaviour does not reproduce.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-35173) Debezium Custom Time Serializer

2024-04-19 Thread ZhengYu Chen (Jira)
ZhengYu Chen created FLINK-35173:


 Summary: Debezium Custom Time Serializer 
 Key: FLINK-35173
 URL: https://issues.apache.org/jira/browse/FLINK-35173
 Project: Flink
  Issue Type: Improvement
  Components: Flink CDC
Affects Versions: 3.1.0
Reporter: ZhengYu Chen
 Fix For: 3.1.0


Currently, Flink CDC Time encounters time type errors (including DateTime, 
Time, Date, TimeStamp) when using MySQL Connector 
(JsonDebeziumDeserializationSchema) as deserialization, and the converted time 
is wrong. The essential reason is that the timestamp returned by the bottom 
layer of debezium is UTC (such as io.debezium.time.Timestamp). The community 
has already had some 
[PR|https://github.com/apache/flink-cdc/pull/1366/files#diff-e129e9fae3eea0bb32f0019debb4932413c91088d6dae656e2ecb63913badae4],
 but they are not work.

Now a way is provided to provide a solution based on Debezium's custom Convert 
interface 
(https://debezium.io/documentation/reference/1.9/development/converters.html),
Users can choose to convert the above four time types into STRING according to 
the specified time format to ensure that users can correctly convert JSON when 
using the Flink DataStream API.


When the user enables this converter, we need to configure it according to the 
parameters, That's some datastream use case:
{code:java}
Properties debeziumProperties = new Properties();
debeziumProperties.setProperty("converters", "datetime");
debeziumProperties.setProperty("datetime.database.type", 
DataBaseType.MYSQL.getType());
debeziumProperties.setProperty("datetime.type", 
"cn.xxx.sources.cdc.MysqlDebeziumConverter");
debeziumProperties.setProperty("datetime.format.date", "-MM-dd");
debeziumProperties.setProperty("datetime.format.time", "HH:mm:ss");
debeziumProperties.setProperty("datetime.format.datetime", "-MM-dd 
HH:mm:ss");
debeziumProperties.setProperty("datetime.format.timestamp", "-MM-dd 
HH:mm:ss");
debeziumProperties.setProperty("datetime.format.timestamp.zone", "UTC+8");
MySqlSourceBuilder builder = MySqlSource.builder()
        .hostname(url[0])
        .port(Integer.parseInt(url[1]))
        .databaseList(table.getDatabase())
        .tableList(getTablePattern(table))
        .username(table.getUserName())
        .password(table.getPassword())
        .debeziumProperties(debeziumProperties); {code}
 

 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-35028][runtime] Timer firing under async execution model [flink]

2024-04-19 Thread via GitHub


fredia commented on PR #24672:
URL: https://github.com/apache/flink/pull/24672#issuecomment-2066225062

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-35156][Runtime] Make operators of DataStream V2 integrate with async state processing framework [flink]

2024-04-19 Thread via GitHub


reswqa commented on code in PR #24678:
URL: https://github.com/apache/flink/pull/24678#discussion_r1572117471


##
flink-datastream/src/main/java/org/apache/flink/datastream/impl/operators/ProcessOperator.java:
##
@@ -23,15 +23,15 @@
 import org.apache.flink.datastream.impl.common.TimestampCollector;
 import org.apache.flink.datastream.impl.context.DefaultNonPartitionedContext;
 import org.apache.flink.datastream.impl.context.DefaultRuntimeContext;
-import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator;
 import org.apache.flink.streaming.api.operators.BoundedOneInput;
 import org.apache.flink.streaming.api.operators.ChainingStrategy;
 import org.apache.flink.streaming.api.operators.OneInputStreamOperator;
+import 
org.apache.flink.streaming.runtime.operators.asyncprocessing.AbstractAsyncStateUdfStreamOperator;
 import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
 
 /** Operator for {@link OneInputStreamProcessFunction}. */
 public class ProcessOperator
-extends AbstractUdfStreamOperator>
+extends AbstractAsyncStateUdfStreamOperator>

Review Comment:
   I think we should also do this for 
`TwoInputBroadcastProcessOperator`/`TwoInputNonBroadcastProcessOperator` and 
`TwoOutputProcessOperator` 樂



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-35115) Kinesis connector writes wrong Kinesis sequence number at stop with savepoint

2024-04-19 Thread Danny Cranmer (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-35115?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17838934#comment-17838934
 ] 

Danny Cranmer commented on FLINK-35115:
---

Merged commit 
[{{8d29147}}|https://github.com/apache/flink-connector-aws/commit/8d29147b9e6c0a7d27399662c6023ad634363764]
 into apache:main 

> Kinesis connector writes wrong Kinesis sequence number at stop with savepoint
> -
>
> Key: FLINK-35115
> URL: https://issues.apache.org/jira/browse/FLINK-35115
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kinesis
>Affects Versions: 1.15.4, aws-connector-4.0.0, aws-connector-4.1.0, 
> aws-connector-4.2.0, 1.16.3, 1.17.2, 1.18.1
> Environment: The issue happens in a *Kinesis -> Flink -> Kafka* 
> exactly-once setup with:
>  * Flink versions checked 1.16.3 and 1.18.1
>  * Kinesis connector checked 1.16.3 and 4.2.0-1.18
>  * checkpointing configured at 1 minute with EXACTLY_ONCE mode: 
> {code:java}
> StreamExecutionEnvironment execEnv = 
> StreamExecutionEnvironment.getExecutionEnvironment (); 
> execEnv.enableCheckpointing (6,EXACTLY_ONCE); execEnv.getCheckpointConfig 
> ().setCheckpointTimeout (9); execEnv.getCheckpointConfig 
> ().setCheckpointStorage (CHECKPOINTS_PATH); {code}
>  * Kafka sink configured with EXACTLY_ONCE semantic/delivery guarantee:
> {code:java}
> Properties sinkConfig = new Properties ();
> sinkConfig.put ("transaction.timeout.ms", 48);
> KafkaSink sink = KafkaSink.builder ()
> .setBootstrapServers ("localhost:9092")
> .setTransactionalIdPrefix ("test-prefix")
> .setDeliverGuarantee (EXACTLY_ONCE)
> .setKafkaProducerConfig (sinkConfig)
> .setRecordSerializer (
> (KafkaRecordSerializationSchema) (element, context, 
> timestamp) -> new ProducerRecord<> (
> "test-output-topic", null, element.getBytes ()))
> .build (); {code}
>  * Kinesis consumer defined as: 
> {code:java}
> FlinkKinesisConsumer flinkKinesisConsumer = new
> FlinkKinesisConsumer<> ("test-stream",
> new AbstractDeserializationSchema<> () {
> @Override
> public ByteBuffer deserialize (byte[] bytes) {
> // Return
> return ByteBuffer.wrap (bytes);
> }
> }, props); {code}
>  
>Reporter: Vadim Vararu
>Assignee: Aleksandr Pilipenko
>Priority: Blocker
>  Labels: kinesis, pull-request-available
> Fix For: aws-connector-4.3.0
>
>
> Having an exactly-once Kinesis -> Flink -> Kafka job and triggering a 
> stop-with-savepoint, Flink duplicates in Kafka all the records between the 
> last checkpoint and the savepoint at resume:
>  * Event1 is written to Kinesis
>  * Event1 is processed by Flink 
>  * Event1 is committed to Kafka at the checkpoint
>  * 
> 
>  * Event2 is written to Kinesis
>  * Event2 is processed by Flink
>  * Stop with savepoint is triggered manually
>  * Event2 is committed to Kafka
>  * 
> 
>  * Job is resumed from the savepoint
>  * *{color:#FF}Event2 is written again to Kafka at the first 
> checkpoint{color}*
>  
> {color:#172b4d}I believe that it's a Kinesis connector issue for 2 
> reasons:{color}
>  * I've checked the actual Kinesis sequence number in the _metadata file 
> generated at stop-with-savepoint and it's the one from the checkpoint before 
> the savepoint  instead of being the one of the last record committed to Kafka.
>  * I've tested exactly the save job with Kafka as source instead of Kinesis 
> as source and the behaviour does not reproduce.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   >