Re: [PR] [FLINK-34593][release] Add release note for version 1.19 [flink]

2024-03-17 Thread via GitHub


lincoln-lil merged PR #24505:
URL: https://github.com/apache/flink/pull/24505


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Add GPG key for 1.19.0 release [flink-docker]

2024-03-17 Thread via GitHub


lincoln-lil merged PR #187:
URL: https://github.com/apache/flink-docker/pull/187


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527940518


##
flink-process-function-parent/flink-process-function/src/main/java/org/apache/flink/process/impl/ExecutionEnvironmentImpl.java:
##
@@ -0,0 +1,341 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl;
+
+import org.apache.flink.api.common.ExecutionConfig;
+import org.apache.flink.api.common.RuntimeExecutionMode;
+import org.apache.flink.api.common.eventtime.WatermarkStrategy;
+import org.apache.flink.api.common.functions.InvalidTypesException;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.connector.v2.FromDataSource;
+import org.apache.flink.api.connector.v2.Source;
+import org.apache.flink.api.connector.v2.WrappedSource;
+import org.apache.flink.api.dag.Transformation;
+import org.apache.flink.api.java.typeutils.MissingTypeInfo;
+import org.apache.flink.api.java.typeutils.ResultTypeQueryable;
+import org.apache.flink.api.java.typeutils.TypeExtractor;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.DeploymentOptions;
+import org.apache.flink.configuration.ExecutionOptions;
+import org.apache.flink.configuration.ReadableConfig;
+import 
org.apache.flink.connector.datagen.functions.FromElementsGeneratorFunction;
+import org.apache.flink.connector.datagen.source.DataGeneratorSource;
+import org.apache.flink.core.execution.DefaultExecutorServiceLoader;
+import org.apache.flink.core.execution.JobClient;
+import org.apache.flink.core.execution.PipelineExecutor;
+import org.apache.flink.core.execution.PipelineExecutorFactory;
+import org.apache.flink.core.execution.PipelineExecutorServiceLoader;
+import org.apache.flink.process.api.ExecutionEnvironment;
+import org.apache.flink.process.api.stream.NonKeyedPartitionStream;
+import org.apache.flink.process.impl.stream.NonKeyedPartitionStreamImpl;
+import org.apache.flink.streaming.api.environment.CheckpointConfig;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamGraphGenerator;
+import org.apache.flink.streaming.api.transformations.SourceTransformation;
+import 
org.apache.flink.streaming.runtime.translators.ProcessFunctionSinkTransformationTranslator;
+import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.FlinkException;
+import org.apache.flink.util.Preconditions;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.List;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.ExecutionException;
+
+import static 
org.apache.flink.streaming.api.graph.StreamGraphGenerator.DEFAULT_TIME_CHARACTERISTIC;
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/** The implementation of {@link ExecutionEnvironment}. */
+public class ExecutionEnvironmentImpl implements ExecutionEnvironment {
+private final List> transformations = new ArrayList<>();
+
+private final ExecutionConfig executionConfig;
+
+/** Settings that control the checkpointing behavior. */
+private final CheckpointConfig checkpointCfg;
+
+private final Configuration configuration;
+
+private final ClassLoader userClassloader;
+
+private final PipelineExecutorServiceLoader executorServiceLoader;
+
+static {
+try {
+// All transformation translator must be put to a map in 
StreamGraphGenerator, but
+// streaming-java is not depend on process-function module, using 
reflect to handle
+// this.
+
ProcessFunctionSinkTransformationTranslator.registerSinkTransformationTranslator();
+} catch (Exception e) {
+throw new RuntimeException(
+"Can not register process function transformation 
translator.");
+}
+}
+
+/**
+ * The environment of the context (local by default, cluster if invoked 
through command line).
+ */
+private static ExecutionEnvironmentFactory contextEnvironmentFactory = 
null;
+
+public static ExecutionEnvironment newInstance() {
+if (contextEnvironmentFactory != null) {
+

Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527936405


##
flink-process-function-parent/flink-process-function/src/test/java/org/apache/flink/process/impl/ExecutionEnvironmentImplTest.java:
##
@@ -0,0 +1,84 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl;
+
+import org.apache.flink.api.common.typeinfo.Types;
+import org.apache.flink.api.connector.v2.SourceUtils;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.core.execution.DefaultExecutorServiceLoader;
+import org.apache.flink.process.api.ExecutionEnvironment;
+import org.apache.flink.process.api.stream.NonKeyedPartitionStream;
+import org.apache.flink.process.impl.stream.StreamTestUtils;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamNode;
+
+import org.junit.jupiter.api.Test;
+
+import java.util.Arrays;
+import java.util.Collection;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link ExecutionEnvironmentImpl}. */
+class ExecutionEnvironmentImplTest {

Review Comment:
   I have added a new test in `ExecutionEnvironmentImplTest`.
   
   As for the testing of `execute`: It's a bit special, or hard to test with a 
unit test. I will introduce integration tests through other PR, which will be 
covered.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527936405


##
flink-process-function-parent/flink-process-function/src/test/java/org/apache/flink/process/impl/ExecutionEnvironmentImplTest.java:
##
@@ -0,0 +1,84 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl;
+
+import org.apache.flink.api.common.typeinfo.Types;
+import org.apache.flink.api.connector.v2.SourceUtils;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.core.execution.DefaultExecutorServiceLoader;
+import org.apache.flink.process.api.ExecutionEnvironment;
+import org.apache.flink.process.api.stream.NonKeyedPartitionStream;
+import org.apache.flink.process.impl.stream.StreamTestUtils;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamNode;
+
+import org.junit.jupiter.api.Test;
+
+import java.util.Arrays;
+import java.util.Collection;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link ExecutionEnvironmentImpl}. */
+class ExecutionEnvironmentImplTest {

Review Comment:
   I have added a new test in `ExecutionEnvironmentImplTest`.
   
   As for the testing of `execute`: It's a bit special, or hard to test with a 
unit test. I will introduce integration tests through other PR, which will be 
covered in due course.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34713) Updates the docs stable version

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34713:

Summary: Updates the docs stable version  (was: CLONE - Updates the docs 
stable version)

> Updates the docs stable version
> ---
>
> Key: FLINK-34713
> URL: https://issues.apache.org/jira/browse/FLINK-34713
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Priority: Major
>
> Update docs to "stable" in {{docs/config.toml}} in the branch of the 
> _just-released_ version:
>  * Change V{{{}ersion{}}} from {{{}x.y-SNAPSHOT }}to \{{{}x.y.z{}}}, i.e. 
> {{1.6-SNAPSHOT}} to {{1.6.0}}
>  * Change V{{{}ersionTitle{}}} from {{x.y-SNAPSHOT}} to {{{}x.y{}}}, i.e. 
> {{1.6-SNAPSHOT}} to {{1.6}}
>  * Change Branch from {{master}} to {{{}release-x.y{}}}, i.e. {{master}} to 
> {{release-1.6}}
>  * Change {{baseURL}} from 
> {{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
>  to 
> {{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
>  * Change {{javadocs_baseurl}} from 
> {{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
>  to 
> {{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
>  * Change {{IsStable}} to {{true}}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34714) Start End of Life discussion thread for now outdated Flink minor version

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34714:

Summary: Start End of Life discussion thread for now outdated Flink minor 
version  (was: CLONE - Start End of Life discussion thread for now outdated 
Flink minor version)

> Start End of Life discussion thread for now outdated Flink minor version
> 
>
> Key: FLINK-34714
> URL: https://issues.apache.org/jira/browse/FLINK-34714
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Priority: Major
>
> The idea is to discuss whether we should do a final release for the now not 
> supported minor version in the community. Such a minor release shouldn't be 
> covered by the current minor version release managers. Their only 
> responsibility is to trigger the discussion.
> The intention of a final patch release for the now unsupported Flink minor 
> version is to flush out all the fixes that didn't end up in the previous 
> release.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34714) Start End of Life discussion thread for now outdated Flink minor version

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34714?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34714:

Reporter: Lincoln Lee  (was: Matthias Pohl)

> Start End of Life discussion thread for now outdated Flink minor version
> 
>
> Key: FLINK-34714
> URL: https://issues.apache.org/jira/browse/FLINK-34714
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Priority: Major
>
> The idea is to discuss whether we should do a final release for the now not 
> supported minor version in the community. Such a minor release shouldn't be 
> covered by the current minor version release managers. Their only 
> responsibility is to trigger the discussion.
> The intention of a final patch release for the now unsupported Flink minor 
> version is to flush out all the fixes that didn't end up in the previous 
> release.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34713) Updates the docs stable version

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34713:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> Updates the docs stable version
> ---
>
> Key: FLINK-34713
> URL: https://issues.apache.org/jira/browse/FLINK-34713
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Priority: Major
>
> Update docs to "stable" in {{docs/config.toml}} in the branch of the 
> _just-released_ version:
>  * Change V{{{}ersion{}}} from {{{}x.y-SNAPSHOT }}to \{{{}x.y.z{}}}, i.e. 
> {{1.6-SNAPSHOT}} to {{1.6.0}}
>  * Change V{{{}ersionTitle{}}} from {{x.y-SNAPSHOT}} to {{{}x.y{}}}, i.e. 
> {{1.6-SNAPSHOT}} to {{1.6}}
>  * Change Branch from {{master}} to {{{}release-x.y{}}}, i.e. {{master}} to 
> {{release-1.6}}
>  * Change {{baseURL}} from 
> {{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
>  to 
> {{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
>  * Change {{javadocs_baseurl}} from 
> {{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
>  to 
> {{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
>  * Change {{IsStable}} to {{true}}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34711) Other announcements

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34711:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> Other announcements
> ---
>
> Key: FLINK-34711
> URL: https://issues.apache.org/jira/browse/FLINK-34711
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Priority: Major
>
> h3. Recordkeeping
> Use [reporter.apache.org|https://reporter.apache.org/addrelease.html?flink] 
> to seed the information about the release into future project reports.
> (Note: Only PMC members have access report releases. If you do not have 
> access, ask on the mailing list for assistance.)
> h3. Flink blog
> Major or otherwise important releases should have a blog post. Write one if 
> needed for this particular release. Minor releases that don’t introduce new 
> major functionality don’t necessarily need to be blogged (see [flink-web PR 
> #581 for Flink 1.15.3|https://github.com/apache/flink-web/pull/581] as an 
> example for a minor release blog post).
> Please make sure that the release notes of the documentation (see section 
> "Review and update documentation") are linked from the blog post of a major 
> release.
> We usually include the names of all contributors in the announcement blog 
> post. Use the following command to get the list of contributors:
> {code}
> # first line is required to make sort first with uppercase and then lower
> export LC_ALL=C
> export FLINK_PREVIOUS_RELEASE_BRANCH=
> export FLINK_CURRENT_RELEASE_BRANCH=
> # e.g.
> # export FLINK_PREVIOUS_RELEASE_BRANCH=release-1.17
> # export FLINK_CURRENT_RELEASE_BRANCH=release-1.18
> git log $(git merge-base master $FLINK_PREVIOUS_RELEASE_BRANCH)..$(git 
> show-ref --hash ${FLINK_CURRENT_RELEASE_BRANCH}) --pretty=format:"%an%n%cn" | 
> sort  -u | paste -sd, | sed "s/\,/\, /g"
> {code}
> h3. Social media
> Tweet, post on Facebook, LinkedIn, and other platforms. Ask other 
> contributors to do the same.
> h3. Flink Release Wiki page
> Add a summary of things that went well or that went not so well during the 
> release process. This can include feedback from contributors but also more 
> generic things like the release have taken longer than initially anticipated 
> (and why) to give a bit of context to the release process.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34711) Other announcements

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34711:

Summary: Other announcements  (was: CLONE - Other announcements)

> Other announcements
> ---
>
> Key: FLINK-34711
> URL: https://issues.apache.org/jira/browse/FLINK-34711
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Priority: Major
>
> h3. Recordkeeping
> Use [reporter.apache.org|https://reporter.apache.org/addrelease.html?flink] 
> to seed the information about the release into future project reports.
> (Note: Only PMC members have access report releases. If you do not have 
> access, ask on the mailing list for assistance.)
> h3. Flink blog
> Major or otherwise important releases should have a blog post. Write one if 
> needed for this particular release. Minor releases that don’t introduce new 
> major functionality don’t necessarily need to be blogged (see [flink-web PR 
> #581 for Flink 1.15.3|https://github.com/apache/flink-web/pull/581] as an 
> example for a minor release blog post).
> Please make sure that the release notes of the documentation (see section 
> "Review and update documentation") are linked from the blog post of a major 
> release.
> We usually include the names of all contributors in the announcement blog 
> post. Use the following command to get the list of contributors:
> {code}
> # first line is required to make sort first with uppercase and then lower
> export LC_ALL=C
> export FLINK_PREVIOUS_RELEASE_BRANCH=
> export FLINK_CURRENT_RELEASE_BRANCH=
> # e.g.
> # export FLINK_PREVIOUS_RELEASE_BRANCH=release-1.17
> # export FLINK_CURRENT_RELEASE_BRANCH=release-1.18
> git log $(git merge-base master $FLINK_PREVIOUS_RELEASE_BRANCH)..$(git 
> show-ref --hash ${FLINK_CURRENT_RELEASE_BRANCH}) --pretty=format:"%an%n%cn" | 
> sort  -u | paste -sd, | sed "s/\,/\, /g"
> {code}
> h3. Social media
> Tweet, post on Facebook, LinkedIn, and other platforms. Ask other 
> contributors to do the same.
> h3. Flink Release Wiki page
> Add a summary of things that went well or that went not so well during the 
> release process. This can include feedback from contributors but also more 
> generic things like the release have taken longer than initially anticipated 
> (and why) to give a bit of context to the release process.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-34687) Home Page of Flink CDC Documentation

2024-03-17 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu resolved FLINK-34687.

Resolution: Resolved

Resolved in flink-cdc(master) via : aa9e91ce4599d8c5a51df82cea4e31aed7d2f634

> Home Page of Flink CDC Documentation
> 
>
> Key: FLINK-34687
> URL: https://issues.apache.org/jira/browse/FLINK-34687
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Qingsheng Ren
>Assignee: Leonard Xu
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.1.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34687][cdc][docs] Build the home page of Flink CDC documentation [flink-cdc]

2024-03-17 Thread via GitHub


leonardBang merged PR #3156:
URL: https://github.com/apache/flink-cdc/pull/3156


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34648] add waitChangeResultRequest and WaitChangeResultResponse to avoid RPC timeout. [flink-cdc]

2024-03-17 Thread via GitHub


BaoPiao commented on PR #3128:
URL: https://github.com/apache/flink-cdc/pull/3128#issuecomment-2003029727

   > @BaoPiao Yeah, this is because SchemaRegistryRequestHandler doesn't clear 
pendingSchemaChanges when SchemaOperator restarted, I've added one commit 
[7aba6a4](https://github.com/apache/flink-cdc/commit/7aba6a4489486af9fece954f561dca0c9c443006)
 to fixed it.
   
   Should we clear the flushedSinkWriters as well?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34710) Apache mailing lists announcements

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34710?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34710:

Summary: Apache mailing lists announcements  (was: CLONE - Apache mailing 
lists announcements)

> Apache mailing lists announcements
> --
>
> Key: FLINK-34710
> URL: https://issues.apache.org/jira/browse/FLINK-34710
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Priority: Major
>
> Announce on the {{dev@}} mailing list that the release has been finished.
> Announce on the release on the {{user@}} mailing list, listing major 
> improvements and contributions.
> Announce the release on the [annou...@apache.org|mailto:annou...@apache.org] 
> mailing list.
> {panel}
> {panel}
> |{{From: Release Manager}}
> {{To: d...@flink.apache.org, u...@flink.apache.org, user...@flink.apache.org, 
> annou...@apache.org}}
> {{Subject: [ANNOUNCE] Apache Flink 1.2.3 released}}
>  
> {{The Apache Flink community is very happy to announce the release of Apache 
> Flink 1.2.3, which is the third bugfix release for the Apache Flink 1.2 
> series.}}
>  
> {{Apache Flink® is an open-source stream processing framework for 
> distributed, high-performing, always-available, and accurate data streaming 
> applications.}}
>  
> {{The release is available for download at:}}
> {{[https://flink.apache.org/downloads.html]}}
>  
> {{Please check out the release blog post for an overview of the improvements 
> for this bugfix release:}}
> {{}}
>  
> {{The full release notes are available in Jira:}}
> {{}}
>  
> {{We would like to thank all contributors of the Apache Flink community who 
> made this release possible!}}
>  
> {{Feel free to reach out to the release managers (or respond to this thread) 
> with feedback on the release process. Our goal is to constantly improve the 
> release process. Feedback on what could be improved or things that didn't go 
> so well are appreciated.}}
>  
> {{Regards,}}
> {{Release Manager}}|



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34710) Apache mailing lists announcements

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34710?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34710:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> Apache mailing lists announcements
> --
>
> Key: FLINK-34710
> URL: https://issues.apache.org/jira/browse/FLINK-34710
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Priority: Major
>
> Announce on the {{dev@}} mailing list that the release has been finished.
> Announce on the release on the {{user@}} mailing list, listing major 
> improvements and contributions.
> Announce the release on the [annou...@apache.org|mailto:annou...@apache.org] 
> mailing list.
> {panel}
> {panel}
> |{{From: Release Manager}}
> {{To: d...@flink.apache.org, u...@flink.apache.org, user...@flink.apache.org, 
> annou...@apache.org}}
> {{Subject: [ANNOUNCE] Apache Flink 1.2.3 released}}
>  
> {{The Apache Flink community is very happy to announce the release of Apache 
> Flink 1.2.3, which is the third bugfix release for the Apache Flink 1.2 
> series.}}
>  
> {{Apache Flink® is an open-source stream processing framework for 
> distributed, high-performing, always-available, and accurate data streaming 
> applications.}}
>  
> {{The release is available for download at:}}
> {{[https://flink.apache.org/downloads.html]}}
>  
> {{Please check out the release blog post for an overview of the improvements 
> for this bugfix release:}}
> {{}}
>  
> {{The full release notes are available in Jira:}}
> {{}}
>  
> {{We would like to thank all contributors of the Apache Flink community who 
> made this release possible!}}
>  
> {{Feel free to reach out to the release managers (or respond to this thread) 
> with feedback on the release process. Our goal is to constantly improve the 
> release process. Feedback on what could be improved or things that didn't go 
> so well are appreciated.}}
>  
> {{Regards,}}
> {{Release Manager}}|



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34708) Merge website pull request

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34708:

Summary: Merge website pull request  (was: CLONE - Merge website pull 
request)

> Merge website pull request
> --
>
> Key: FLINK-34708
> URL: https://issues.apache.org/jira/browse/FLINK-34708
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>
> Merge the website pull request to [list the 
> release|http://flink.apache.org/downloads.html]. Make sure to regenerate the 
> website as well, as it isn't build automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34709) Remove outdated versions

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34709:

Summary: Remove outdated versions  (was: CLONE - Remove outdated versions)

> Remove outdated versions
> 
>
> Key: FLINK-34709
> URL: https://issues.apache.org/jira/browse/FLINK-34709
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Priority: Major
>
> h4. dist.apache.org
> For a new major release remove all release files older than 2 versions, e.g., 
> when releasing 1.7, remove all releases <= 1.5.
> For a new bugfix version remove all release files for previous bugfix 
> releases in the same series, e.g., when releasing 1.7.1, remove the 1.7.0 
> release.
> # If you have not already, check out the Flink section of the {{release}} 
> repository on {{[dist.apache.org|http://dist.apache.org/]}} via Subversion. 
> In a fresh directory:
> {code}
> svn checkout https://dist.apache.org/repos/dist/release/flink 
> --depth=immediates
> cd flink
> {code}
> # Remove files for outdated releases and commit the changes.
> {code}
> svn remove flink-
> svn commit
> {code}
> # Verify that files  are 
> [removed|https://dist.apache.org/repos/dist/release/flink]
> (!) Remember to remove the corresponding download links from the website.
> h4. CI
> Disable the cron job for the now-unsupported version from 
> (tools/azure-pipelines/[build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
>  in the respective branch.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-34712) CLONE - Update reference data for Migration Tests

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee reassigned FLINK-34712:
---

Assignee: lincoln lee  (was: Sergey Nuyanzin)

> CLONE - Update reference data for Migration Tests
> -
>
> Key: FLINK-34712
> URL: https://issues.apache.org/jira/browse/FLINK-34712
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update migration tests in master to cover migration from new version. Since 
> 1.18, this step could be done automatically with the following steps. For 
> more information please refer to [this 
> page.|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/README.md]
>  # {*}On the published release tag (e.g., release-1.16.0){*}, run 
> {panel}
> {panel}
> |{{$ mvn clean }}{{package}} {{{}-Pgenerate-migration-test-data 
> -Dgenerate.version={}}}{{{}1.16{}}} {{-nsu -Dfast -DskipTests}}|
> The version (1.16 in the command above) should be replaced with the target 
> one.
>  # Modify the content of the file 
> [apache/flink:flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version]
>  to the latest version (it would be "v1_16" if sticking to the example where 
> 1.16.0 was released). 
>  # Commit the modification in step a and b with "{_}[release] Generate 
> reference data for state migration tests based on release-1.xx.0{_}" to the 
> corresponding release branch (e.g. {{release-1.16}} in our example), replace 
> "xx" with the actual version (in this example "16"). You should use the Jira 
> issue ID in case of [release]  as the commit message's prefix if you have a 
> dedicated Jira issue for this task.
>  # Cherry-pick the commit to the master branch. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34712) CLONE - Update reference data for Migration Tests

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34712:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> CLONE - Update reference data for Migration Tests
> -
>
> Key: FLINK-34712
> URL: https://issues.apache.org/jira/browse/FLINK-34712
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update migration tests in master to cover migration from new version. Since 
> 1.18, this step could be done automatically with the following steps. For 
> more information please refer to [this 
> page.|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/README.md]
>  # {*}On the published release tag (e.g., release-1.16.0){*}, run 
> {panel}
> {panel}
> |{{$ mvn clean }}{{package}} {{{}-Pgenerate-migration-test-data 
> -Dgenerate.version={}}}{{{}1.16{}}} {{-nsu -Dfast -DskipTests}}|
> The version (1.16 in the command above) should be replaced with the target 
> one.
>  # Modify the content of the file 
> [apache/flink:flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version]
>  to the latest version (it would be "v1_16" if sticking to the example where 
> 1.16.0 was released). 
>  # Commit the modification in step a and b with "{_}[release] Generate 
> reference data for state migration tests based on release-1.xx.0{_}" to the 
> corresponding release branch (e.g. {{release-1.16}} in our example), replace 
> "xx" with the actual version (in this example "16"). You should use the Jira 
> issue ID in case of [release]  as the commit message's prefix if you have a 
> dedicated Jira issue for this task.
>  # Cherry-pick the commit to the master branch. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34712) CLONE - Update reference data for Migration Tests

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34712:

Fix Version/s: 1.20.0
   1.19.1
   (was: 1.19.0)
   (was: 1.18.1)

> CLONE - Update reference data for Migration Tests
> -
>
> Key: FLINK-34712
> URL: https://issues.apache.org/jira/browse/FLINK-34712
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0, 1.19.1
>
>
> Update migration tests in master to cover migration from new version. Since 
> 1.18, this step could be done automatically with the following steps. For 
> more information please refer to [this 
> page.|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/README.md]
>  # {*}On the published release tag (e.g., release-1.16.0){*}, run 
> {panel}
> {panel}
> |{{$ mvn clean }}{{package}} {{{}-Pgenerate-migration-test-data 
> -Dgenerate.version={}}}{{{}1.16{}}} {{-nsu -Dfast -DskipTests}}|
> The version (1.16 in the command above) should be replaced with the target 
> one.
>  # Modify the content of the file 
> [apache/flink:flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version]
>  to the latest version (it would be "v1_16" if sticking to the example where 
> 1.16.0 was released). 
>  # Commit the modification in step a and b with "{_}[release] Generate 
> reference data for state migration tests based on release-1.xx.0{_}" to the 
> corresponding release branch (e.g. {{release-1.16}} in our example), replace 
> "xx" with the actual version (in this example "16"). You should use the Jira 
> issue ID in case of [release]  as the commit message's prefix if you have a 
> dedicated Jira issue for this task.
>  # Cherry-pick the commit to the master branch. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34712) Update reference data for Migration Tests

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34712:

Summary: Update reference data for Migration Tests  (was: CLONE - Update 
reference data for Migration Tests)

> Update reference data for Migration Tests
> -
>
> Key: FLINK-34712
> URL: https://issues.apache.org/jira/browse/FLINK-34712
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0, 1.19.1
>
>
> Update migration tests in master to cover migration from new version. Since 
> 1.18, this step could be done automatically with the following steps. For 
> more information please refer to [this 
> page.|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/README.md]
>  # {*}On the published release tag (e.g., release-1.16.0){*}, run 
> {panel}
> {panel}
> |{{$ mvn clean }}{{package}} {{{}-Pgenerate-migration-test-data 
> -Dgenerate.version={}}}{{{}1.16{}}} {{-nsu -Dfast -DskipTests}}|
> The version (1.16 in the command above) should be replaced with the target 
> one.
>  # Modify the content of the file 
> [apache/flink:flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version]
>  to the latest version (it would be "v1_16" if sticking to the example where 
> 1.16.0 was released). 
>  # Commit the modification in step a and b with "{_}[release] Generate 
> reference data for state migration tests based on release-1.xx.0{_}" to the 
> corresponding release branch (e.g. {{release-1.16}} in our example), replace 
> "xx" with the actual version (in this example "16"). You should use the Jira 
> issue ID in case of [release]  as the commit message's prefix if you have a 
> dedicated Jira issue for this task.
>  # Cherry-pick the commit to the master branch. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34709) Remove outdated versions

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34709:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> Remove outdated versions
> 
>
> Key: FLINK-34709
> URL: https://issues.apache.org/jira/browse/FLINK-34709
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Priority: Major
>
> h4. dist.apache.org
> For a new major release remove all release files older than 2 versions, e.g., 
> when releasing 1.7, remove all releases <= 1.5.
> For a new bugfix version remove all release files for previous bugfix 
> releases in the same series, e.g., when releasing 1.7.1, remove the 1.7.0 
> release.
> # If you have not already, check out the Flink section of the {{release}} 
> repository on {{[dist.apache.org|http://dist.apache.org/]}} via Subversion. 
> In a fresh directory:
> {code}
> svn checkout https://dist.apache.org/repos/dist/release/flink 
> --depth=immediates
> cd flink
> {code}
> # Remove files for outdated releases and commit the changes.
> {code}
> svn remove flink-
> svn commit
> {code}
> # Verify that files  are 
> [removed|https://dist.apache.org/repos/dist/release/flink]
> (!) Remember to remove the corresponding download links from the website.
> h4. CI
> Disable the cron job for the now-unsupported version from 
> (tools/azure-pipelines/[build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
>  in the respective branch.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34707) CLONE - Update japicmp configuration

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34707:

Fix Version/s: 1.20.0
   1.19.1
   (was: 1.19.0)
   (was: 1.18.1)

> CLONE - Update japicmp configuration
> 
>
> Key: FLINK-34707
> URL: https://issues.apache.org/jira/browse/FLINK-34707
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0, 1.19.1
>
>
> Update the japicmp reference version and wipe exclusions / enable API 
> compatibility checks for {{@PublicEvolving}} APIs on the corresponding 
> SNAPSHOT branch with the {{update_japicmp_configuration.sh}} script (see 
> below).
> For a new major release (x.y.0), run the same command also on the master 
> branch for updating the japicmp reference version and removing out-dated 
> exclusions in the japicmp configuration.
> Make sure that all Maven artifacts are already pushed to Maven Central. 
> Otherwise, there's a risk that CI fails due to missing reference artifacts.
> {code:bash}
> tools $ NEW_VERSION=$RELEASE_VERSION releasing/update_japicmp_configuration.sh
> tools $ cd ..$ git add *$ git commit -m "Update japicmp configuration for 
> $RELEASE_VERSION" {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-34708) CLONE - Merge website pull request

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee reassigned FLINK-34708:
---

Assignee: lincoln lee

> CLONE - Merge website pull request
> --
>
> Key: FLINK-34708
> URL: https://issues.apache.org/jira/browse/FLINK-34708
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>
> Merge the website pull request to [list the 
> release|http://flink.apache.org/downloads.html]. Make sure to regenerate the 
> website as well, as it isn't build automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34707) Update japicmp configuration

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34707:

Summary: Update japicmp configuration  (was: CLONE - Update japicmp 
configuration)

> Update japicmp configuration
> 
>
> Key: FLINK-34707
> URL: https://issues.apache.org/jira/browse/FLINK-34707
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0, 1.19.1
>
>
> Update the japicmp reference version and wipe exclusions / enable API 
> compatibility checks for {{@PublicEvolving}} APIs on the corresponding 
> SNAPSHOT branch with the {{update_japicmp_configuration.sh}} script (see 
> below).
> For a new major release (x.y.0), run the same command also on the master 
> branch for updating the japicmp reference version and removing out-dated 
> exclusions in the japicmp configuration.
> Make sure that all Maven artifacts are already pushed to Maven Central. 
> Otherwise, there's a risk that CI fails due to missing reference artifacts.
> {code:bash}
> tools $ NEW_VERSION=$RELEASE_VERSION releasing/update_japicmp_configuration.sh
> tools $ cd ..$ git add *$ git commit -m "Update japicmp configuration for 
> $RELEASE_VERSION" {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34708) CLONE - Merge website pull request

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34708:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> CLONE - Merge website pull request
> --
>
> Key: FLINK-34708
> URL: https://issues.apache.org/jira/browse/FLINK-34708
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>
> Merge the website pull request to [list the 
> release|http://flink.apache.org/downloads.html]. Make sure to regenerate the 
> website as well, as it isn't build automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34648] add waitChangeResultRequest and WaitChangeResultResponse to avoid RPC timeout. [flink-cdc]

2024-03-17 Thread via GitHub


yanghuaiGit commented on PR #3128:
URL: https://github.com/apache/flink-cdc/pull/3128#issuecomment-2003023705

   
如果schemaRegister的pending里有多个待执行schema事件,isSchemaChangeApplying代表的是正在执行schema变更事件,但是无法得知是哪个schema事件执行,isSchemaChangeApplying无法和客户端的事件绑定,导致schemaoperator可能需要等待所有的schema事件执行完才会转为非阻塞


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-34707) CLONE - Update japicmp configuration

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee reassigned FLINK-34707:
---

Assignee: lincoln lee  (was: Sergey Nuyanzin)

> CLONE - Update japicmp configuration
> 
>
> Key: FLINK-34707
> URL: https://issues.apache.org/jira/browse/FLINK-34707
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update the japicmp reference version and wipe exclusions / enable API 
> compatibility checks for {{@PublicEvolving}} APIs on the corresponding 
> SNAPSHOT branch with the {{update_japicmp_configuration.sh}} script (see 
> below).
> For a new major release (x.y.0), run the same command also on the master 
> branch for updating the japicmp reference version and removing out-dated 
> exclusions in the japicmp configuration.
> Make sure that all Maven artifacts are already pushed to Maven Central. 
> Otherwise, there's a risk that CI fails due to missing reference artifacts.
> {code:bash}
> tools $ NEW_VERSION=$RELEASE_VERSION releasing/update_japicmp_configuration.sh
> tools $ cd ..$ git add *$ git commit -m "Update japicmp configuration for 
> $RELEASE_VERSION" {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34714) CLONE - Start End of Life discussion thread for now outdated Flink minor version

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34714:
---

 Summary: CLONE - Start End of Life discussion thread for now 
outdated Flink minor version
 Key: FLINK-34714
 URL: https://issues.apache.org/jira/browse/FLINK-34714
 Project: Flink
  Issue Type: Sub-task
Reporter: Matthias Pohl


The idea is to discuss whether we should do a final release for the now not 
supported minor version in the community. Such a minor release shouldn't be 
covered by the current minor version release managers. Their only 
responsibility is to trigger the discussion.

The intention of a final patch release for the now unsupported Flink minor 
version is to flush out all the fixes that didn't end up in the previous 
release.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34711) CLONE - Other announcements

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34711:
---

 Summary: CLONE - Other announcements
 Key: FLINK-34711
 URL: https://issues.apache.org/jira/browse/FLINK-34711
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin


h3. Recordkeeping

Use [reporter.apache.org|https://reporter.apache.org/addrelease.html?flink] to 
seed the information about the release into future project reports.

(Note: Only PMC members have access report releases. If you do not have access, 
ask on the mailing list for assistance.)
h3. Flink blog

Major or otherwise important releases should have a blog post. Write one if 
needed for this particular release. Minor releases that don’t introduce new 
major functionality don’t necessarily need to be blogged (see [flink-web PR 
#581 for Flink 1.15.3|https://github.com/apache/flink-web/pull/581] as an 
example for a minor release blog post).

Please make sure that the release notes of the documentation (see section 
"Review and update documentation") are linked from the blog post of a major 
release.
We usually include the names of all contributors in the announcement blog post. 
Use the following command to get the list of contributors:
{code}
# first line is required to make sort first with uppercase and then lower
export LC_ALL=C
export FLINK_PREVIOUS_RELEASE_BRANCH=
export FLINK_CURRENT_RELEASE_BRANCH=
# e.g.
# export FLINK_PREVIOUS_RELEASE_BRANCH=release-1.17
# export FLINK_CURRENT_RELEASE_BRANCH=release-1.18
git log $(git merge-base master $FLINK_PREVIOUS_RELEASE_BRANCH)..$(git show-ref 
--hash ${FLINK_CURRENT_RELEASE_BRANCH}) --pretty=format:"%an%n%cn" | sort  -u | 
paste -sd, | sed "s/\,/\, /g"
{code}
h3. Social media

Tweet, post on Facebook, LinkedIn, and other platforms. Ask other contributors 
to do the same.
h3. Flink Release Wiki page

Add a summary of things that went well or that went not so well during the 
release process. This can include feedback from contributors but also more 
generic things like the release have taken longer than initially anticipated 
(and why) to give a bit of context to the release process.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-34706) Promote release 1.19

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee reassigned FLINK-34706:
---

Assignee: lincoln lee  (was: Jing Ge)

> Promote release 1.19
> 
>
> Key: FLINK-34706
> URL: https://issues.apache.org/jira/browse/FLINK-34706
> Project: Flink
>  Issue Type: New Feature
>Affects Versions: 1.18.0
>Reporter: Sergey Nuyanzin
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
>
> Once the release has been finalized (FLINK-32920), the last step of the 
> process is to promote the release within the project and beyond. Please wait 
> for 24h after finalizing the release in accordance with the [ASF release 
> policy|http://www.apache.org/legal/release-policy.html#release-announcements].
> *Final checklist to declare this issue resolved:*
>  # Website pull request to [list the 
> release|http://flink.apache.org/downloads.html] merged
>  # Release announced on the user@ mailing list.
>  # Blog post published, if applicable.
>  # Release recorded in 
> [reporter.apache.org|https://reporter.apache.org/addrelease.html?flink].
>  # Release announced on social media.
>  # Completion declared on the dev@ mailing list.
>  # Update Homebrew: 
> [https://docs.brew.sh/How-To-Open-a-Homebrew-Pull-Request] (seems to be done 
> automatically - at least for minor releases  for both minor and major 
> releases)
>  # Updated the japicmp configuration
>  ** corresponding SNAPSHOT branch japicmp reference version set to the just 
> released version, and API compatibiltity checks for {{@PublicEvolving}}  was 
> enabled
>  ** (minor version release only) master branch japicmp reference version set 
> to the just released version
>  ** (minor version release only) master branch japicmp exclusions have been 
> cleared
>  # Update the list of previous version in {{docs/config.toml}} on the master 
> branch.
>  # Set {{show_outdated_warning: true}} in {{docs/config.toml}} in the branch 
> of the _now deprecated_ Flink version (i.e. 1.16 if 1.18.0 is released)
>  # Update stable and master alias in 
> [https://github.com/apache/flink/blob/master/.github/workflows/docs.yml]
>  # Open discussion thread for End of Life for Unsupported version (i.e. 1.16)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34706) Promote release 1.19

2024-03-17 Thread lincoln lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

lincoln lee updated FLINK-34706:

Reporter: Lincoln Lee  (was: Sergey Nuyanzin)

> Promote release 1.19
> 
>
> Key: FLINK-34706
> URL: https://issues.apache.org/jira/browse/FLINK-34706
> Project: Flink
>  Issue Type: New Feature
>Affects Versions: 1.18.0
>Reporter: Lincoln Lee
>Assignee: lincoln lee
>Priority: Major
>  Labels: pull-request-available
>
> Once the release has been finalized (FLINK-32920), the last step of the 
> process is to promote the release within the project and beyond. Please wait 
> for 24h after finalizing the release in accordance with the [ASF release 
> policy|http://www.apache.org/legal/release-policy.html#release-announcements].
> *Final checklist to declare this issue resolved:*
>  # Website pull request to [list the 
> release|http://flink.apache.org/downloads.html] merged
>  # Release announced on the user@ mailing list.
>  # Blog post published, if applicable.
>  # Release recorded in 
> [reporter.apache.org|https://reporter.apache.org/addrelease.html?flink].
>  # Release announced on social media.
>  # Completion declared on the dev@ mailing list.
>  # Update Homebrew: 
> [https://docs.brew.sh/How-To-Open-a-Homebrew-Pull-Request] (seems to be done 
> automatically - at least for minor releases  for both minor and major 
> releases)
>  # Updated the japicmp configuration
>  ** corresponding SNAPSHOT branch japicmp reference version set to the just 
> released version, and API compatibiltity checks for {{@PublicEvolving}}  was 
> enabled
>  ** (minor version release only) master branch japicmp reference version set 
> to the just released version
>  ** (minor version release only) master branch japicmp exclusions have been 
> cleared
>  # Update the list of previous version in {{docs/config.toml}} on the master 
> branch.
>  # Set {{show_outdated_warning: true}} in {{docs/config.toml}} in the branch 
> of the _now deprecated_ Flink version (i.e. 1.16 if 1.18.0 is released)
>  # Update stable and master alias in 
> [https://github.com/apache/flink/blob/master/.github/workflows/docs.yml]
>  # Open discussion thread for End of Life for Unsupported version (i.e. 1.16)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34713) CLONE - Updates the docs stable version

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34713:
---

 Summary: CLONE - Updates the docs stable version
 Key: FLINK-34713
 URL: https://issues.apache.org/jira/browse/FLINK-34713
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin


Update docs to "stable" in {{docs/config.toml}} in the branch of the 
_just-released_ version:
 * Change V{{{}ersion{}}} from {{{}x.y-SNAPSHOT }}to \{{{}x.y.z{}}}, i.e. 
{{1.6-SNAPSHOT}} to {{1.6.0}}
 * Change V{{{}ersionTitle{}}} from {{x.y-SNAPSHOT}} to {{{}x.y{}}}, i.e. 
{{1.6-SNAPSHOT}} to {{1.6}}
 * Change Branch from {{master}} to {{{}release-x.y{}}}, i.e. {{master}} to 
{{release-1.6}}
 * Change {{baseURL}} from 
{{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
 to 
{{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
 * Change {{javadocs_baseurl}} from 
{{//[ci.apache.org/projects/flink/flink-docs-master|http://ci.apache.org/projects/flink/flink-docs-master]}}
 to 
{{//[ci.apache.org/projects/flink/flink-docs-release-x.y|http://ci.apache.org/projects/flink/flink-docs-release-x.y]}}
 * Change {{IsStable}} to {{true}}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34712) CLONE - Update reference data for Migration Tests

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34712:
---

 Summary: CLONE - Update reference data for Migration Tests
 Key: FLINK-34712
 URL: https://issues.apache.org/jira/browse/FLINK-34712
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin
Assignee: Sergey Nuyanzin
 Fix For: 1.19.0, 1.18.1


Update migration tests in master to cover migration from new version. Since 
1.18, this step could be done automatically with the following steps. For more 
information please refer to [this 
page.|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/README.md]
 # {*}On the published release tag (e.g., release-1.16.0){*}, run 
{panel}
{panel}
|{{$ mvn clean }}{{package}} {{{}-Pgenerate-migration-test-data 
-Dgenerate.version={}}}{{{}1.16{}}} {{-nsu -Dfast -DskipTests}}|

The version (1.16 in the command above) should be replaced with the target one.

 # Modify the content of the file 
[apache/flink:flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version|https://github.com/apache/flink/blob/master/flink-test-utils-parent/flink-migration-test-utils/src/main/resources/most_recently_published_version]
 to the latest version (it would be "v1_16" if sticking to the example where 
1.16.0 was released). 
 # Commit the modification in step a and b with "{_}[release] Generate 
reference data for state migration tests based on release-1.xx.0{_}" to the 
corresponding release branch (e.g. {{release-1.16}} in our example), replace 
"xx" with the actual version (in this example "16"). You should use the Jira 
issue ID in case of [release]  as the commit message's prefix if you have a 
dedicated Jira issue for this task.

 # Cherry-pick the commit to the master branch. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34710) CLONE - Apache mailing lists announcements

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34710:
---

 Summary: CLONE - Apache mailing lists announcements
 Key: FLINK-34710
 URL: https://issues.apache.org/jira/browse/FLINK-34710
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin


Announce on the {{dev@}} mailing list that the release has been finished.

Announce on the release on the {{user@}} mailing list, listing major 
improvements and contributions.

Announce the release on the [annou...@apache.org|mailto:annou...@apache.org] 
mailing list.
{panel}
{panel}
|{{From: Release Manager}}
{{To: d...@flink.apache.org, u...@flink.apache.org, user...@flink.apache.org, 
annou...@apache.org}}
{{Subject: [ANNOUNCE] Apache Flink 1.2.3 released}}
 
{{The Apache Flink community is very happy to announce the release of Apache 
Flink 1.2.3, which is the third bugfix release for the Apache Flink 1.2 
series.}}
 
{{Apache Flink® is an open-source stream processing framework for distributed, 
high-performing, always-available, and accurate data streaming applications.}}
 
{{The release is available for download at:}}
{{[https://flink.apache.org/downloads.html]}}
 
{{Please check out the release blog post for an overview of the improvements 
for this bugfix release:}}
{{}}
 
{{The full release notes are available in Jira:}}
{{}}
 
{{We would like to thank all contributors of the Apache Flink community who 
made this release possible!}}
 
{{Feel free to reach out to the release managers (or respond to this thread) 
with feedback on the release process. Our goal is to constantly improve the 
release process. Feedback on what could be improved or things that didn't go so 
well are appreciated.}}
 
{{Regards,}}
{{Release Manager}}|



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34709) CLONE - Remove outdated versions

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34709:
---

 Summary: CLONE - Remove outdated versions
 Key: FLINK-34709
 URL: https://issues.apache.org/jira/browse/FLINK-34709
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin


h4. dist.apache.org

For a new major release remove all release files older than 2 versions, e.g., 
when releasing 1.7, remove all releases <= 1.5.

For a new bugfix version remove all release files for previous bugfix releases 
in the same series, e.g., when releasing 1.7.1, remove the 1.7.0 release.
# If you have not already, check out the Flink section of the {{release}} 
repository on {{[dist.apache.org|http://dist.apache.org/]}} via Subversion. In 
a fresh directory:
{code}
svn checkout https://dist.apache.org/repos/dist/release/flink --depth=immediates
cd flink
{code}
# Remove files for outdated releases and commit the changes.
{code}
svn remove flink-
svn commit
{code}
# Verify that files  are 
[removed|https://dist.apache.org/repos/dist/release/flink]
(!) Remember to remove the corresponding download links from the website.

h4. CI

Disable the cron job for the now-unsupported version from 
(tools/azure-pipelines/[build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 in the respective branch.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34708) CLONE - Merge website pull request

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34708:
---

 Summary: CLONE - Merge website pull request
 Key: FLINK-34708
 URL: https://issues.apache.org/jira/browse/FLINK-34708
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin


Merge the website pull request to [list the 
release|http://flink.apache.org/downloads.html]. Make sure to regenerate the 
website as well, as it isn't build automatically.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34707) CLONE - Update japicmp configuration

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34707:
---

 Summary: CLONE - Update japicmp configuration
 Key: FLINK-34707
 URL: https://issues.apache.org/jira/browse/FLINK-34707
 Project: Flink
  Issue Type: Sub-task
Reporter: Sergey Nuyanzin
Assignee: Sergey Nuyanzin
 Fix For: 1.19.0, 1.18.1


Update the japicmp reference version and wipe exclusions / enable API 
compatibility checks for {{@PublicEvolving}} APIs on the corresponding SNAPSHOT 
branch with the {{update_japicmp_configuration.sh}} script (see below).

For a new major release (x.y.0), run the same command also on the master branch 
for updating the japicmp reference version and removing out-dated exclusions in 
the japicmp configuration.

Make sure that all Maven artifacts are already pushed to Maven Central. 
Otherwise, there's a risk that CI fails due to missing reference artifacts.
{code:bash}
tools $ NEW_VERSION=$RELEASE_VERSION releasing/update_japicmp_configuration.sh
tools $ cd ..$ git add *$ git commit -m "Update japicmp configuration for 
$RELEASE_VERSION" {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34706) Promote release 1.19

2024-03-17 Thread lincoln lee (Jira)
lincoln lee created FLINK-34706:
---

 Summary: Promote release 1.19
 Key: FLINK-34706
 URL: https://issues.apache.org/jira/browse/FLINK-34706
 Project: Flink
  Issue Type: New Feature
Affects Versions: 1.18.0
Reporter: Sergey Nuyanzin
Assignee: Jing Ge


Once the release has been finalized (FLINK-32920), the last step of the process 
is to promote the release within the project and beyond. Please wait for 24h 
after finalizing the release in accordance with the [ASF release 
policy|http://www.apache.org/legal/release-policy.html#release-announcements].

*Final checklist to declare this issue resolved:*
 # Website pull request to [list the 
release|http://flink.apache.org/downloads.html] merged
 # Release announced on the user@ mailing list.
 # Blog post published, if applicable.
 # Release recorded in 
[reporter.apache.org|https://reporter.apache.org/addrelease.html?flink].
 # Release announced on social media.
 # Completion declared on the dev@ mailing list.
 # Update Homebrew: [https://docs.brew.sh/How-To-Open-a-Homebrew-Pull-Request] 
(seems to be done automatically - at least for minor releases  for both minor 
and major releases)
 # Updated the japicmp configuration
 ** corresponding SNAPSHOT branch japicmp reference version set to the just 
released version, and API compatibiltity checks for {{@PublicEvolving}}  was 
enabled
 ** (minor version release only) master branch japicmp reference version set to 
the just released version
 ** (minor version release only) master branch japicmp exclusions have been 
cleared
 # Update the list of previous version in {{docs/config.toml}} on the master 
branch.
 # Set {{show_outdated_warning: true}} in {{docs/config.toml}} in the branch of 
the _now deprecated_ Flink version (i.e. 1.16 if 1.18.0 is released)
 # Update stable and master alias in 
[https://github.com/apache/flink/blob/master/.github/workflows/docs.yml]
 # Open discussion thread for End of Life for Unsupported version (i.e. 1.16)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [build] fix jackson conflicts among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


Shawn-Hx commented on code in PR #2987:
URL: https://github.com/apache/flink-cdc/pull/2987#discussion_r1527909731


##
pom.xml:
##
@@ -95,6 +96,36 @@ under the License.
 
 
 
+
+com.fasterxml.jackson.core

Review Comment:
   Thanks for your contribution. I think using 
[jackson-bom](https://github.com/FasterXML/jackson-bom) here is more clear.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34668][checkpoint] Report operator state handle of file merging directory to JM [flink]

2024-03-17 Thread via GitHub


flinkbot commented on PR #24513:
URL: https://github.com/apache/flink/pull/24513#issuecomment-2003011938

   
   ## CI report:
   
   * 8cc6814c6ea3b5dee6280f350ec293cb06ece658 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34684 & 34683][docs] Add licenses & contributing docs in Flink CDC [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on code in PR #3159:
URL: https://github.com/apache/flink-cdc/pull/3159#discussion_r1527773536


##
docs/content/docs/developer-guide/contribute-to-flink-cdc.md:
##
@@ -23,3 +23,62 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Contributing
+
+Flink CDC is developed by an open and friendly community and welcomes anyone 
who wants to help out in any way.
+There are several ways to interact with the community and contribute to Flink 
CDC including asking questions, filing 
+bug reports, proposing new features, joining discussions on the mailing lists, 
contributing code or documentation, 
+improving website, testing release candidates and writing corresponding blog 
etc.
+
+## What do you want to do
+
+Contributing to Flink CDC goes beyond writing code for the project. Here are 
different opportunities to help the 
+project as follows.
+
+| Area| Further information


   |
+|:|:--|
+| Report Bug  | To report a problem, open an issue in [Flink 
jira](https://issues.apache.org/jira/projects/FLINK/issues) and select `Flink 
CDC` in `Component/s`. Please give detailed information about the problem you 
encountered and, if possible, add a description that helps to reproduce the 
problem. |
+| Contribute Code | Read the Code 
Contribution Guide  

 |
+| Code Reviews| Read the Code Review 
Guide   

|

Review Comment:
   ~~We should update the actual page addresses of these two links.~~



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34668) Report State handle of file merging directory to JM

2024-03-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34668?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34668:
---
Labels: pull-request-available  (was: )

> Report State handle of file merging directory to JM
> ---
>
> Key: FLINK-34668
> URL: https://issues.apache.org/jira/browse/FLINK-34668
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing
>Reporter: Zakelly Lan
>Assignee: Yanfei Lei
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-34668][checkpoint] Report operator state handle of file merging directory to JM [flink]

2024-03-17 Thread via GitHub


fredia opened a new pull request, #24513:
URL: https://github.com/apache/flink/pull/24513

   
   
   ## What is the purpose of the change
   
   Operator states are stored in `taskownd/` dir when file merging is enabled, 
this PR reports the operator state handle of file merging directory to JM 
`SharedStateRegistry`. 
   
   ## Brief change log
 - Introduce `FileMergingOperatorStreamStateHandle` , 
`EmptyFileMergingOperatorStreamStateHandle` and `DirectoryStreamStateHandle`.
 - Still report file merging directory to JM even if operator state is 
empty.
   
   ## Verifying this change
   
   This change added tests and can be verified as follows:
   - `OperatorStateBackendTest#testFileMergingSnapshotEmpty`
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (no)
 - The serializers: (no)
 - The runtime per-record code paths (performance sensitive): (no)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: ( checkpointing )
 - The S3 file system connector: ( no )
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527893607


##
flink-process-function-parent/flink-process-function-api/src/main/java/org/apache/flink/process/api/stream/KeyedPartitionStream.java:
##
@@ -0,0 +1,186 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.api.stream;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.api.connector.v2.Sink;
+import org.apache.flink.api.java.functions.KeySelector;
+import org.apache.flink.process.api.function.OneInputStreamProcessFunction;
+import 
org.apache.flink.process.api.function.TwoInputBroadcastStreamProcessFunction;
+import 
org.apache.flink.process.api.function.TwoInputNonBroadcastStreamProcessFunction;
+import org.apache.flink.process.api.function.TwoOutputStreamProcessFunction;
+import 
org.apache.flink.process.api.stream.NonKeyedPartitionStream.TwoNonKeyedPartitionStreams;
+
+/** This interface represents a stream that each parallel task processes the 
same data. */

Review Comment:
   Ai, good catch!.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [build] fix inconsistent Kafka shading among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


link3280 commented on code in PR #2988:
URL: https://github.com/apache/flink-cdc/pull/2988#discussion_r1527893588


##
flink-cdc-connect/flink-cdc-source-connectors/flink-sql-connector-tidb-cdc/pom.xml:
##
@@ -60,6 +60,12 @@ under the License.
 
 
 
+
+org.apache.kafka
+
+
com.ververica.cdc.connectors.shaded.org.apache.kafka

Review Comment:
   I found no shaded pattern has migrated from 
`com.ververica.cdc.connectors.shaded` yet. Maybe we should do it all together 
in an other PR?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527888923


##
flink-process-function-parent/flink-process-function-api/src/main/java/org/apache/flink/process/api/stream/BroadcastStream.java:
##
@@ -0,0 +1,68 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.api.stream;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.api.java.functions.KeySelector;
+import 
org.apache.flink.process.api.function.TwoInputBroadcastStreamProcessFunction;
+
+/** This interface represents a stream that each parallel task processes the 
same data. */
+@Experimental
+public interface BroadcastStream {
+/**
+ * Apply a two input operation to this and other {@link 
KeyedPartitionStream}.
+ *
+ * @param other {@link KeyedPartitionStream} to perform operation with two 
input.
+ * @param processFunction to perform operation.
+ * @return new stream with this operation.
+ */
+ NonKeyedPartitionStream connectAndProcess(
+KeyedPartitionStream other,
+TwoInputBroadcastStreamProcessFunction 
processFunction);
+
+/**
+ * Apply a two input operation to this and other {@link 
NonKeyedPartitionStream}.
+ *
+ * @param other {@link NonKeyedPartitionStream} to perform operation with 
two input.
+ * @param processFunction to perform operation.
+ * @return new stream with this operation.
+ */
+ NonKeyedPartitionStream connectAndProcess(
+NonKeyedPartitionStream other,
+TwoInputBroadcastStreamProcessFunction 
processFunction);
+
+/**
+ * Apply a two input operation to this and other {@link 
KeyedPartitionStream}.
+ *
+ * This method is used to avoid shuffle after applying the process 
function. It is required
+ * that for the record from non-broadcast input, the new {@link 
KeySelector} must extract the
+ * same key as the original {@link KeySelector}s on the {@link 
KeyedPartitionStream}. For the
+ * record from broadcast input, the output key from keyed partition itself 
instead of new key
+ * selector, so it is safe already.

Review Comment:
   Yes, the clearer the better. But I'm not sure what is not clear, could you 
give me some proposed changes?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [build] fix inconsistent Kafka shading among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


link3280 commented on code in PR #2988:
URL: https://github.com/apache/flink-cdc/pull/2988#discussion_r1527886955


##
flink-cdc-connect/flink-cdc-source-connectors/flink-sql-connector-tidb-cdc/pom.xml:
##
@@ -60,6 +60,12 @@ under the License.
 
 
 
+
+org.apache.kafka
+
+
com.ververica.cdc.connectors.shaded.org.apache.kafka

Review Comment:
   You're right. I didn't notice the package name had changed. Hold on a sec.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527883895


##
flink-core-api/pom.xml:
##
@@ -0,0 +1,60 @@
+
+
+http://maven.apache.org/POM/4.0.0"; 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd";>
+
+   4.0.0
+
+   
+   org.apache.flink
+   flink-parent
+   1.20-SNAPSHOT
+   
+
+   flink-core-api
+   Flink : Core API
+
+   jar
+
+   
+   
+   org.apache.flink
+   flink-annotations
+   ${project.version}
+   
+   
+
+   
+   
+   

Review Comment:
   Make sense. I first introduced it and thinking that this PR-moved class from 
flink-core would have a corresponding test class, but it doesn't. :)



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527879616


##
flink-process-function-parent/flink-process-function/src/main/java/org/apache/flink/process/impl/ExecutionEnvironmentImpl.java:
##
@@ -0,0 +1,341 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl;
+
+import org.apache.flink.api.common.ExecutionConfig;
+import org.apache.flink.api.common.RuntimeExecutionMode;
+import org.apache.flink.api.common.eventtime.WatermarkStrategy;
+import org.apache.flink.api.common.functions.InvalidTypesException;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.connector.v2.FromDataSource;
+import org.apache.flink.api.connector.v2.Source;
+import org.apache.flink.api.connector.v2.WrappedSource;
+import org.apache.flink.api.dag.Transformation;
+import org.apache.flink.api.java.typeutils.MissingTypeInfo;
+import org.apache.flink.api.java.typeutils.ResultTypeQueryable;
+import org.apache.flink.api.java.typeutils.TypeExtractor;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.DeploymentOptions;
+import org.apache.flink.configuration.ExecutionOptions;
+import org.apache.flink.configuration.ReadableConfig;
+import 
org.apache.flink.connector.datagen.functions.FromElementsGeneratorFunction;
+import org.apache.flink.connector.datagen.source.DataGeneratorSource;
+import org.apache.flink.core.execution.DefaultExecutorServiceLoader;
+import org.apache.flink.core.execution.JobClient;
+import org.apache.flink.core.execution.PipelineExecutor;
+import org.apache.flink.core.execution.PipelineExecutorFactory;
+import org.apache.flink.core.execution.PipelineExecutorServiceLoader;
+import org.apache.flink.process.api.ExecutionEnvironment;
+import org.apache.flink.process.api.stream.NonKeyedPartitionStream;
+import org.apache.flink.process.impl.stream.NonKeyedPartitionStreamImpl;
+import org.apache.flink.streaming.api.environment.CheckpointConfig;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamGraphGenerator;
+import org.apache.flink.streaming.api.transformations.SourceTransformation;
+import 
org.apache.flink.streaming.runtime.translators.ProcessFunctionSinkTransformationTranslator;
+import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.FlinkException;
+import org.apache.flink.util.Preconditions;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.List;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.ExecutionException;
+
+import static 
org.apache.flink.streaming.api.graph.StreamGraphGenerator.DEFAULT_TIME_CHARACTERISTIC;
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/** The implementation of {@link ExecutionEnvironment}. */
+public class ExecutionEnvironmentImpl implements ExecutionEnvironment {
+private final List> transformations = new ArrayList<>();
+
+private final ExecutionConfig executionConfig;
+
+/** Settings that control the checkpointing behavior. */
+private final CheckpointConfig checkpointCfg;
+
+private final Configuration configuration;
+
+private final ClassLoader userClassloader;
+
+private final PipelineExecutorServiceLoader executorServiceLoader;
+
+static {
+try {
+// All transformation translator must be put to a map in 
StreamGraphGenerator, but
+// streaming-java is not depend on process-function module, using 
reflect to handle
+// this.
+
ProcessFunctionSinkTransformationTranslator.registerSinkTransformationTranslator();
+} catch (Exception e) {
+throw new RuntimeException(
+"Can not register process function transformation 
translator.");
+}
+}
+
+/**
+ * The environment of the context (local by default, cluster if invoked 
through command line).
+ */
+private static ExecutionEnvironmentFactory contextEnvironmentFactory = 
null;
+
+public static ExecutionEnvironment newInstance() {
+if (contextEnvironmentFactory != null) {
+

Re: [PR] [build] fix inconsistent Kafka shading among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


leonardBang commented on code in PR #2988:
URL: https://github.com/apache/flink-cdc/pull/2988#discussion_r1527879357


##
flink-cdc-connect/flink-cdc-source-connectors/flink-sql-connector-tidb-cdc/pom.xml:
##
@@ -60,6 +60,12 @@ under the License.
 
 
 
+
+org.apache.kafka
+
+
com.ververica.cdc.connectors.shaded.org.apache.kafka

Review Comment:
   Maybe we should change `  com.ververica.`  to `org.apache.flink` after we 
donated to Apache?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34679][cdc] add doc under core-concept. [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on PR #3153:
URL: https://github.com/apache/flink-cdc/pull/3153#issuecomment-2002977044

   address it.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527870916


##
flink-process-function-parent/flink-process-function/src/main/java/org/apache/flink/streaming/api/transformations/ProcessFunctionSinkTransformation.java:
##
@@ -0,0 +1,96 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.transformations;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.connector.sink2.Sink;
+import org.apache.flink.api.connector.sink2.SinkWriter;
+import org.apache.flink.api.dag.Transformation;
+import org.apache.flink.process.impl.stream.DataStream;
+import org.apache.flink.streaming.api.operators.ChainingStrategy;
+
+import org.apache.commons.compress.utils.Lists;
+
+import javax.annotation.Nullable;
+
+import java.util.Collections;
+import java.util.List;
+
+import static org.apache.flink.util.Preconditions.checkNotNull;
+
+/**
+ * A {@link Transformation} for sink-v2 in process function.
+ *
+ * @param  The input type of the {@link SinkWriter}
+ * @param  The output type of the {@link Sink}
+ */
+@Internal
+public class ProcessFunctionSinkTransformation

Review Comment:
   `SinkTransformation` and `DataStream(v1 version)` are tightly coupled. 
Unfortunately, V1 and V2 can't be converted to each other.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527867946


##
flink-process-function-parent/flink-process-function/src/main/java/org/apache/flink/process/impl/operators/ProcessOperator.java:
##
@@ -0,0 +1,71 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl.operators;
+
+import org.apache.flink.process.api.function.OneInputStreamProcessFunction;
+import org.apache.flink.process.impl.common.OutputCollector;
+import org.apache.flink.process.impl.common.TimestampCollector;
+import org.apache.flink.process.impl.context.DefaultNonPartitionedContext;
+import org.apache.flink.process.impl.context.DefaultRuntimeContext;
+import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator;
+import org.apache.flink.streaming.api.operators.BoundedOneInput;
+import org.apache.flink.streaming.api.operators.ChainingStrategy;
+import org.apache.flink.streaming.api.operators.OneInputStreamOperator;
+import org.apache.flink.streaming.runtime.streamrecord.StreamRecord;
+
+/** Operator for {@link OneInputStreamProcessFunction}. */
+public class ProcessOperator
+extends AbstractUdfStreamOperator>
+implements OneInputStreamOperator, BoundedOneInput {
+
+protected transient DefaultRuntimeContext context;
+
+protected transient DefaultNonPartitionedContext 
nonPartitionedContext;
+
+protected transient TimestampCollector outputCollector;
+
+public ProcessOperator(OneInputStreamProcessFunction 
userFunction) {
+super(userFunction);
+
+chainingStrategy = ChainingStrategy.ALWAYS;

Review Comment:
   
https://github.com/apache/flink/blob/841f23c73e4399df91112dd11ddca74f45ea5b37/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/operators/SimpleOperatorFactory.java#L92-L98
   
   IIRC, `AbstractStreamOperator` is always a `SetupableStreamOperator` with 
the default chainingStrategy `HEAD`. 🤔 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34689][MySQL][Feature] check binlog_row_value_optoins [flink-cdc]

2024-03-17 Thread via GitHub


leonardBang commented on PR #3148:
URL: https://github.com/apache/flink-cdc/pull/3148#issuecomment-2002970708

   Thanks @SML0127 for the contribution, @ruanhang1993 Would you like to help 
review this PR?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


reswqa commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1527863818


##
flink-process-function-parent/flink-process-function-api/src/main/java/org/apache/flink/process/api/function/ProcessFunction.java:
##
@@ -0,0 +1,50 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.api.function;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.api.common.functions.Function;
+
+/** Base class for all user defined process functions. */
+@Experimental
+public interface ProcessFunction extends Function {
+/**
+ * Initialization method for the function. It is called before the actual 
working methods (like
+ * processRecord) and thus suitable for one time setup work.
+ *
+ * By default, this method does nothing.
+ *
+ * @throws Exception Implementations may forward exceptions, which are 
caught by the runtime.
+ * When the runtime catches an exception, it aborts the task and lets 
the fail-over logic
+ * decide whether to retry the task execution.
+ */
+default void open() throws Exception {}

Review Comment:
   I just didn't realize what we need to do in `OpenContext` at the moment? 
Since the V2 API is experimental, we can introduce such contexts to it before 
it becomes public, I think.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-34702) Rank should not convert to StreamExecDuplicate when the input is not insert only

2024-03-17 Thread lincoln lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34702?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17827843#comment-17827843
 ] 

lincoln lee commented on FLINK-34702:
-

[~jackylau] Thank you for reporting this! 
This should be an improvement but not a bug. Currently, some operators do not 
support processing changelog input, and `StreamPhysicalDeduplicate` is one of 
them.
For the solution, can we consider taking a step forward, allowing 
`StreamPhysicalDeduplicate` to support changelog input (let the physical 
operator support this input, just like we added support changelog input for 
Window TVF Aggregation in 1.19), instead of patching with a switch during the 
optimization phase to bypass it, WDYT?

> Rank should not convert to StreamExecDuplicate when the input is not insert 
> only
> 
>
> Key: FLINK-34702
> URL: https://issues.apache.org/jira/browse/FLINK-34702
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.20.0
>Reporter: Jacky Lau
>Priority: Major
> Fix For: 1.20.0
>
>
> {code:java}
> @Test
> def testSimpleFirstRowOnBuiltinProctime1(): Unit = {
>   val sqlQuery =
> """
>   |SELECT *
>   |FROM (
>   |  SELECT *,
>   |ROW_NUMBER() OVER (PARTITION BY a ORDER BY PROCTIME() ASC) as 
> rowNum
>   |  FROM (select a, count(b) as b from MyTable group by a)
>   |)
>   |WHERE rowNum = 1
> """.stripMargin
>   util.verifyExecPlan(sqlQuery)
> } {code}
> Exception:
> org.apache.flink.table.api.TableException: StreamPhysicalDeduplicate doesn't 
> support consuming update changes which is produced by node 
> GroupAggregate(groupBy=[a], select=[a, COUNT(b) AS b])
> because the StreamPhysicalDeduplicate can not consuming update changes now 
> while StreamExecRank can.
> so we should not convert the FlinkLogicalRank to StreamPhysicalDeduplicate in 
> this case. and we can defer whether input contains update change in the 
> "optimize the physical plan" phase. 
> so we can add an option to solve it. and when the StreamPhysicalDeduplicate 
> can support consuming update changes , we can deprecate it



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-34689) check binlog_row_value_optoins

2024-03-17 Thread Lee SeungMin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34689?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17827840#comment-17827840
 ] 

Lee SeungMin commented on FLINK-34689:
--

Assign this pr to me please

> check binlog_row_value_optoins
> --
>
> Key: FLINK-34689
> URL: https://issues.apache.org/jira/browse/FLINK-34689
> Project: Flink
>  Issue Type: Improvement
>  Components: Flink CDC
>Reporter: Lee SeungMin
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2024-03-15-12-56-49-344.png
>
>
> When {{binlog_row_value_optoins}} is set to {{{}PARTIAL_JSON{}}},
> the update operator remains as {{{}Update_rows_partial{}}}.
> Flink CDC does not parse this event because {{Update_row_partial}} binlog 
> event is mapped to {{PARTIAL_UPDATE_ROWS_EVENT}} and Flink CDC do not handle 
> that event type
>  
> Example of Update_row_partial (when {{binlog_row_value_optoins}} = 
> {{PARTIAL_JSON)}}
> !image-2024-03-15-12-56-49-344.png|width=1015,height=30!
> So, we have to check {{binlog_row_value_optoins}} before starting.
>  
>  
> Cretae PR: [[MySQL][Feature] check binlog_row_value_optoins by SML0127 · Pull 
> Request #3148 · apache/flink-cdc 
> (github.com)|https://github.com/apache/flink-cdc/pull/3148]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34680) "Connectors - Overview" Page for Flink CDC Documentation

2024-03-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34680:
---
Labels: pull-request-available  (was: )

> "Connectors - Overview" Page for  Flink CDC Documentation
> -
>
> Key: FLINK-34680
> URL: https://issues.apache.org/jira/browse/FLINK-34680
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Qingsheng Ren
>Assignee: Qingsheng Ren
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.1.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-34705) org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase FAILURE

2024-03-17 Thread baihailiang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

baihailiang updated FLINK-34705:

Description: 
 
{code:java}
// code placeholder
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 111.197 
s <<< FAILURE! - in 
org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase
[ERROR] org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase 
 Time elapsed: 111.197 s  <<< ERROR!
org.testcontainers.containers.ContainerLaunchException: Container startup failed
    at 
org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:345)
    at 
org.testcontainers.containers.GenericContainer.start(GenericContainer.java:326)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension$StoreAdapter.start(TestcontainersExtension.java:242)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension$StoreAdapter.access$200(TestcontainersExtension.java:229)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.lambda$null$1(TestcontainersExtension.java:59)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore.lambda$getOrComputeIfAbsent$4(ExtensionValuesStore.java:86)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$MemoizingSupplier.computeValue(ExtensionValuesStore.java:223)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$MemoizingSupplier.get(ExtensionValuesStore.java:211)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$StoredValue.evaluate(ExtensionValuesStore.java:191)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$StoredValue.access$100(ExtensionValuesStore.java:171)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore.getOrComputeIfAbsent(ExtensionValuesStore.java:89)
    at 
org.junit.jupiter.engine.execution.NamespaceAwareStore.getOrComputeIfAbsent(NamespaceAwareStore.java:53)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.lambda$beforeAll$2(TestcontainersExtension.java:59)
    at java.util.ArrayList.forEach(ArrayList.java:1259)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.beforeAll(TestcontainersExtension.java:59)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeBeforeAllCallbacks$10(ClassBasedTestDescriptor.java:381)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeBeforeAllCallbacks(ClassBasedTestDescriptor.java:381)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:205)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:80)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:148)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
    at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
    at 
org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185)
    at 
org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService.invokeAll(ForkJoinPoolHierarchicalTestExecutorService.java:129)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
    at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
    at 
org.junit.platform.engine.support.hierarchical.ForkJoinPoolH

[jira] [Updated] (FLINK-34548) FLIP-409: DataStream V2 Building Blocks: DataStream, Partitioning and ProcessFunction

2024-03-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34548:
---
Labels: pull-request-available  (was: )

> FLIP-409: DataStream V2 Building Blocks: DataStream, Partitioning and 
> ProcessFunction
> -
>
> Key: FLINK-34548
> URL: https://issues.apache.org/jira/browse/FLINK-34548
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / DataStream
>Reporter: Weijie Guo
>Assignee: Weijie Guo
>Priority: Major
>  Labels: pull-request-available
>
> This is the ticket for FLIP-409: DataStream V2 Building Blocks: DataStream, 
> Partitioning and ProcessFunction.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-34705) org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase FAILURE

2024-03-17 Thread baihailiang (Jira)
baihailiang created FLINK-34705:
---

 Summary: 
org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase FAILURE
 Key: FLINK-34705
 URL: https://issues.apache.org/jira/browse/FLINK-34705
 Project: Flink
  Issue Type: Bug
  Components: Connectors / ElasticSearch
Affects Versions: 1.16.2
 Environment: Environment is x86 machine
Reporter: baihailiang


 
{code:java}
// code placeholder
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 111.197 
s <<< FAILURE! - in 
org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase
[ERROR] org.apache.flink.connector.elasticsearch.sink.ElasticsearchWriterITCase 
 Time elapsed: 111.197 s  <<< ERROR!
org.testcontainers.containers.ContainerLaunchException: Container startup failed
    at 
org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:345)
    at 
org.testcontainers.containers.GenericContainer.start(GenericContainer.java:326)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension$StoreAdapter.start(TestcontainersExtension.java:242)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension$StoreAdapter.access$200(TestcontainersExtension.java:229)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.lambda$null$1(TestcontainersExtension.java:59)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore.lambda$getOrComputeIfAbsent$4(ExtensionValuesStore.java:86)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$MemoizingSupplier.computeValue(ExtensionValuesStore.java:223)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$MemoizingSupplier.get(ExtensionValuesStore.java:211)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$StoredValue.evaluate(ExtensionValuesStore.java:191)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore$StoredValue.access$100(ExtensionValuesStore.java:171)
    at 
org.junit.jupiter.engine.execution.ExtensionValuesStore.getOrComputeIfAbsent(ExtensionValuesStore.java:89)
    at 
org.junit.jupiter.engine.execution.NamespaceAwareStore.getOrComputeIfAbsent(NamespaceAwareStore.java:53)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.lambda$beforeAll$2(TestcontainersExtension.java:59)
    at java.util.ArrayList.forEach(ArrayList.java:1259)
    at 
org.testcontainers.junit.jupiter.TestcontainersExtension.beforeAll(TestcontainersExtension.java:59)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeBeforeAllCallbacks$10(ClassBasedTestDescriptor.java:381)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeBeforeAllCallbacks(ClassBasedTestDescriptor.java:381)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:205)
    at 
org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:80)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:148)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
    at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95)
    at 
org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService$ExclusiveTask.compute(ForkJoinPoolHierarchicalTestExecutorService.java:185)
    at 
org.junit.platform.engine.support.hierarchical.ForkJoinPoolHierarchicalTestExecutorService.invokeAll(ForkJoinPoolHierarchicalTestExecutorService.java:129)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141)
    at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137)
    at 
org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139)
    at 
org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.jav

Re: [PR] [FLINK-34548][api] Introduce DataStream, Partitioning and ProcessFunction [flink]

2024-03-17 Thread via GitHub


WencongLiu commented on code in PR #24422:
URL: https://github.com/apache/flink/pull/24422#discussion_r1523101423


##
flink-process-function-parent/flink-process-function-api/src/main/java/org/apache/flink/process/api/stream/BroadcastStream.java:
##
@@ -0,0 +1,68 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.api.stream;
+
+import org.apache.flink.annotation.Experimental;
+import org.apache.flink.api.java.functions.KeySelector;
+import 
org.apache.flink.process.api.function.TwoInputBroadcastStreamProcessFunction;
+
+/** This interface represents a stream that each parallel task processes the 
same data. */
+@Experimental
+public interface BroadcastStream {
+/**
+ * Apply a two input operation to this and other {@link 
KeyedPartitionStream}.
+ *
+ * @param other {@link KeyedPartitionStream} to perform operation with two 
input.
+ * @param processFunction to perform operation.
+ * @return new stream with this operation.
+ */
+ NonKeyedPartitionStream connectAndProcess(
+KeyedPartitionStream other,
+TwoInputBroadcastStreamProcessFunction 
processFunction);
+
+/**
+ * Apply a two input operation to this and other {@link 
NonKeyedPartitionStream}.
+ *
+ * @param other {@link NonKeyedPartitionStream} to perform operation with 
two input.
+ * @param processFunction to perform operation.
+ * @return new stream with this operation.
+ */
+ NonKeyedPartitionStream connectAndProcess(
+NonKeyedPartitionStream other,
+TwoInputBroadcastStreamProcessFunction 
processFunction);
+
+/**
+ * Apply a two input operation to this and other {@link 
KeyedPartitionStream}.
+ *
+ * This method is used to avoid shuffle after applying the process 
function. It is required
+ * that for the record from non-broadcast input, the new {@link 
KeySelector} must extract the
+ * same key as the original {@link KeySelector}s on the {@link 
KeyedPartitionStream}. For the
+ * record from broadcast input, the output key from keyed partition itself 
instead of new key
+ * selector, so it is safe already.

Review Comment:
   For `KeyedPartitionStream` and `BroadcastStream`,  we may need to explain 
more details on why we need a new key selector, why it avoid shuffle process, 
and the concept of `safe`.



##
flink-process-function-parent/flink-process-function/src/main/java/org/apache/flink/process/impl/utils/StreamUtils.java:
##
@@ -0,0 +1,285 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.process.impl.utils;
+
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.connector.v2.Sink;
+import org.apache.flink.api.connector.v2.WrappedSink;
+import org.apache.flink.api.java.Utils;
+import org.apache.flink.api.java.functions.KeySelector;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.api.java.typeutils.TypeExtractor;
+import org.apache.flink.process.api.function.OneInputStreamProcessFunction;
+import 
org.apache.flink.process.api.function.TwoInputBroadcastStreamProcessFunction;
+import 
org.apache.flink.process.api.function.TwoInputNonBroadcastStreamProcessFunction;
+import org.apache.flink.process.api.function.TwoOutputStreamProcessFunction;
+import org.apache.flink.process.impl.stream.DataStream;
+import org.apache.flink.pr

Re: [PR] [FLINK-29114][connector][filesystem] Fix issue of file overwriting caused by multiple writes to the same sink table and shared staging directory [flink]

2024-03-17 Thread via GitHub


flinkbot commented on PR #24512:
URL: https://github.com/apache/flink/pull/24512#issuecomment-2002894151

   
   ## CI report:
   
   * 79ca29c5dd966fe4af885c9e5b18a737c5e5b6e9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-29114][connector][filesystem] Fix issue of file overwriting caused by multiple writes to the same sink table and shared staging directory [flink]

2024-03-17 Thread via GitHub


LadyForest opened a new pull request, #24512:
URL: https://github.com/apache/flink/pull/24512

   ## What is the purpose of the change
   
   This PR is cherry-picked from https://github.com/apache/flink/pull/24390
   
   ## Brief change log
   
   - Fix unstable TableSourceITCase#testTableHintWithLogicalTableScanReuse
   - Moves the staging dir configuration into builder for easier testing
   
   ## Verifying this change
   
   FileSystemOutputFormatTest#testGetUniqueStagingDirectory
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34684 & 34683][docs] Add licenses & contributing docs in Flink CDC [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on code in PR #3159:
URL: https://github.com/apache/flink-cdc/pull/3159#discussion_r1527773536


##
docs/content/docs/developer-guide/contribute-to-flink-cdc.md:
##
@@ -23,3 +23,62 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Contributing
+
+Flink CDC is developed by an open and friendly community and welcomes anyone 
who wants to help out in any way.
+There are several ways to interact with the community and contribute to Flink 
CDC including asking questions, filing 
+bug reports, proposing new features, joining discussions on the mailing lists, 
contributing code or documentation, 
+improving website, testing release candidates and writing corresponding blog 
etc.
+
+## What do you want to do
+
+Contributing to Flink CDC goes beyond writing code for the project. Here are 
different opportunities to help the 
+project as follows.
+
+| Area| Further information


   |
+|:|:--|
+| Report Bug  | To report a problem, open an issue in [Flink 
jira](https://issues.apache.org/jira/projects/FLINK/issues) and select `Flink 
CDC` in `Component/s`. Please give detailed information about the problem you 
encountered and, if possible, add a description that helps to reproduce the 
problem. |
+| Contribute Code | Read the Code 
Contribution Guide  

 |
+| Code Reviews| Read the Code Review 
Guide   

|

Review Comment:
   We should update the actual page addresses of these two links.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34684 & 34683][docs] Add licenses & contributing docs in Flink CDC [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on code in PR #3159:
URL: https://github.com/apache/flink-cdc/pull/3159#discussion_r1527773536


##
docs/content/docs/developer-guide/contribute-to-flink-cdc.md:
##
@@ -23,3 +23,62 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Contributing
+
+Flink CDC is developed by an open and friendly community and welcomes anyone 
who wants to help out in any way.
+There are several ways to interact with the community and contribute to Flink 
CDC including asking questions, filing 
+bug reports, proposing new features, joining discussions on the mailing lists, 
contributing code or documentation, 
+improving website, testing release candidates and writing corresponding blog 
etc.
+
+## What do you want to do
+
+Contributing to Flink CDC goes beyond writing code for the project. Here are 
different opportunities to help the 
+project as follows.
+
+| Area| Further information


   |
+|:|:--|
+| Report Bug  | To report a problem, open an issue in [Flink 
jira](https://issues.apache.org/jira/projects/FLINK/issues) and select `Flink 
CDC` in `Component/s`. Please give detailed information about the problem you 
encountered and, if possible, add a description that helps to reproduce the 
problem. |
+| Contribute Code | Read the Code 
Contribution Guide  

 |
+| Code Reviews| Read the Code Review 
Guide   

|

Review Comment:
   We should update the actual addresses of these two links.



##
docs/content/docs/developer-guide/contribute-to-flink-cdc.md:
##
@@ -23,3 +23,62 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Contributing
+
+Flink CDC is developed by an open and friendly community and welcomes anyone 
who wants to help out in any way.
+There are several ways to interact with the community and contribute to Flink 
CDC including asking questions, filing 
+bug reports, proposing new features, joining discussions on the mailing lists, 
contributing code or documentation, 
+improving website, testing release candidates and writing corresponding blog 
etc.
+
+## What do you want to do
+
+Contributing to Flink CDC goes beyond writing code for the project. Here are 
different opportunities to help the 
+project as follows.
+
+| Area| Further information


   |
+|:|:--|
+| Report Bug  | To report a problem, open an issue in [Flink 
jira](https://issues.apache.org/jira/projects/FLINK/issues) and select `Flink 
CDC` in `Component/s`. Please give detailed information about the problem you 
encountered and, if possible, add a description that helps to reproduce the 
problem. |
+| Contribute Code | Read the Code 
Contribution Guide  

 |
+| Code Reviews| Read the Code Review 
Guide   

|
+| Support Users   | Reply to questions on the flink user mailing list, check 
the latest is

Re: [PR] [FLINK-34679][cdc] add doc under core-concept. [flink-cdc]

2024-03-17 Thread via GitHub


leonardBang commented on code in PR #3153:
URL: https://github.com/apache/flink-cdc/pull/3153#discussion_r1527763427


##
docs/content/docs/core-concept/data-pipeline.md:
##
@@ -23,3 +23,80 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Definition
+Since events in Flink CDC flow from the upstream to the downstream in a 
pipeline manner, the data synchronization task is also referred as a Data 
Pipeline.

Review Comment:
   we can use **Data Pipeline** to highlight



##
docs/content/docs/core-concept/route.md:
##
@@ -23,3 +23,52 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Definition
+Route specifies the rule of matching a list of source-table and mapping to 
sink-table. The most typical scenario is the merge of sub-databases and 
sub-tables, routing multiple upstream source tables to the same sink table.
+
+# Parameters
+To describe a route, the follows are required:  
+
+| parameter | meaning |
+|---|--|
+| source-table | Source table id, supports regular expressions   |
+| sink-table | Sink table id, supports regular expressions   |
+| description | Routing rule description(optional, default 
value provided)|
+
+A route module can contain a list of source-table/sink-table rules.
+
+# Example
+## one to one
+if synchronize the table `web_order` in the database `mydb` to a Doris table 
`ods_web_order`, we can use this yaml file to define this route:
+
+```yaml
+route:
+source-table: mydb.web_order
+sink-table: mydb.ods_web_order
+description: sync table to one destination table with given prefix ods_
+```
+
+## many to one

Review Comment:
   Route multiple Data Source tables to one Data Sink table



##
docs/content/docs/core-concept/data-pipeline.md:
##
@@ -23,3 +23,80 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Definition
+Since events in Flink CDC flow from the upstream to the downstream in a 
pipeline manner, the data synchronization task is also referred as a Data 
Pipeline.
+
+# Parameters
+A pipeline corresponds to a chain of operators in Flink.   
+To describe a Data Pipeline, the following parts are required:
+- [source](data-source.md)
+- [sink](data-sink.md)
+- [pipeline](data-pipeline.md#global-parameters) (global parameters)
+
+the following parts are optional:
+- [route](route.md)
+- [transform](transform.md)
+
+# Example
+## Only required
+We could use this concise yaml file to define a pipeline:
+
+```yaml
+   source:
+ type: mysql
+ hostname: localhost
+ port: 3306
+ username: root
+ password: 123456
+ tables: app_db.\.*
+
+   sink:
+ type: doris
+ fenodes: 127.0.0.1:8030
+ username: root
+ password: ""
+
+   pipeline:
+ name: Sync MySQL Database to Doris
+ parallelism: 2
+```
+
+## With optional
+We could use this complicated yaml file to define a pipeline:
+
+```yaml
+   source:
+ type: mysql
+ hostname: localhost
+ port: 3306
+ username: root
+ password: 123456
+ tables: app_db.\.*
+
+   sink:
+ type: doris
+ fenodes: 127.0.0.1:8030
+ username: root
+ password: ""
+   route:
+ - source-table: app_db.orders
+   sink-table: ods_db.ods_orders
+ - source-table: app_db.shipments
+   sink-table: ods_db.ods_shipments
+ - source-table: app_db.products
+   sink-table: ods_db.ods_products  
+
+   pipeline:
+ name: Sync MySQL Database to Doris
+ parallelism: 2
+```
+
+# Global Parameters
+The following parameters are global parameters of the pipeline:

Review Comment:
   The following config options of Data Pipeline level are supported:
   



##
docs/content/docs/core-concept/data-pipeline.md:
##
@@ -23,3 +23,80 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Definition
+Since events in Flink CDC flow from the upstream to the downstream in a 
pipeline manner, the data synchronization task is also referred as a Data 
Pipeline.

Review Comment:
the data synchronization task  -> the whole ETL task is referred as a   ** 
Data Pipeline **



##
docs/content/docs/core-concept/route.md:
##
@@ -23,3 +23,52 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Definition
+Route specifies the rule of matching a list of source-table and mapping to 
sink-table. The most typical scenario is the merge of sub-databases and 
sub-tables, routing multiple upstream source tables to the same sink t

[jira] [Created] (FLINK-34704) Process checkpoint barrier in AsyncWaitOperator when the element queue is full

2024-03-17 Thread Zakelly Lan (Jira)
Zakelly Lan created FLINK-34704:
---

 Summary: Process checkpoint barrier in AsyncWaitOperator when the 
element queue is full
 Key: FLINK-34704
 URL: https://issues.apache.org/jira/browse/FLINK-34704
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Task
Reporter: Zakelly Lan


As discussed in 
https://lists.apache.org/thread/4f7ywn29kdv4302j2rq3fkxc6pc8myr2 . Maybe it is 
better to provide such a new `yield` that can process mail with low priority in 
the mailbox executor. More discussion needed.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-29114][connector][filesystem] Fix issue of file overwriting caused by multiple writes to the same sink table and shared staging directory [flink]

2024-03-17 Thread via GitHub


flinkbot commented on PR #24511:
URL: https://github.com/apache/flink/pull/24511#issuecomment-2002875245

   
   ## CI report:
   
   * 09a83bf383c13f3262a61279871cc0e9882f6e47 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34678][cdc] Add introduction page for Flink CDC docs [flink-cdc]

2024-03-17 Thread via GitHub


leonardBang commented on code in PR #3158:
URL: https://github.com/apache/flink-cdc/pull/3158#discussion_r1527756592


##
docs/content/docs/get-started/introduction.md:
##
@@ -23,3 +23,98 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Welcome to Flink CDC 🎉
+
+Flink CDC is a stream data integration framework that aims to provide users 
with

Review Comment:
   Flink CDC is a streaming data integration tool 



##
docs/content/docs/get-started/introduction.md:
##
@@ -23,3 +23,98 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Welcome to Flink CDC 🎉
+
+Flink CDC is a stream data integration framework that aims to provide users 
with
+a more robust API. It allows users to configure their data synchronization 
logic
+through customized Flink operators and job submission tools. The framework
+prioritizes optimizing the task submission process and offers enhanced
+functionalities such as whole database synchronization, sharding, and schema
+change synchronization.

Review Comment:
   please keep same with homepage key feature section.



##
docs/content/docs/get-started/introduction.md:
##
@@ -23,3 +23,98 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Welcome to Flink CDC 🎉
+
+Flink CDC is a stream data integration framework that aims to provide users 
with
+a more robust API. It allows users to configure their data synchronization 
logic
+through customized Flink operators and job submission tools. The framework
+prioritizes optimizing the task submission process and offers enhanced

Review Comment:
   The framework prioritizes
   > Flink CDC prioritizes



##
docs/content/docs/get-started/introduction.md:
##
@@ -23,3 +23,98 @@ KIND, either express or implied.  See the License for the
 specific language governing permissions and limitations
 under the License.
 -->
+
+# Welcome to Flink CDC 🎉
+
+Flink CDC is a stream data integration framework that aims to provide users 
with
+a more robust API. It allows users to configure their data synchronization 
logic
+through customized Flink operators and job submission tools. The framework

Review Comment:
It allows users to describe their ETL pipeline logic via YAML elegantly and 
help users automatically generating customized Flink operators and submitting 
job.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-29114][connector][filesystem] Fix issue of file overwriting caused by multiple writes to the same sink table and shared staging directory [flink]

2024-03-17 Thread via GitHub


LadyForest opened a new pull request, #24511:
URL: https://github.com/apache/flink/pull/24511

   ## What is the purpose of the change
   
   This PR is cherry-picked from https://github.com/apache/flink/pull/24390
   
   
   ## Brief change log
   
   - Fix unstable TableSourceITCase#testTableHintWithLogicalTableScanReuse
   - Moves the staging dir configuration into builder for easier testing
   
   
   ## Verifying this change
   
   FileSystemOutputFormatTest#testGetUniqueStagingDirectory
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializer: no
 - The runtime per-record code paths (performance sensitive): no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34687][cdc][docs] Build the home page of Flink CDC documentation [flink-cdc]

2024-03-17 Thread via GitHub


PatrickRen commented on code in PR #3156:
URL: https://github.com/apache/flink-cdc/pull/3156#discussion_r1527754771


##
docs/content/_index.md:
##
@@ -22,17 +22,179 @@ specific language governing permissions and limitations
 under the License.
 -->
 
- 
+
+
+
+
+Flink CDC
+
+
+The streaming data integration tool
+
+
+
+
+
+
+
+
+
+
+
+  
+  {{< img src="/fig/cdc-flow.png" alt="Flink CDC Flow">}}
+  
+
+
+
+
+
 
-
-  
-Flink CDC: Change Data Capture Solution Of Apache Flink
-  
-Set of source connectors for Apache Flink® directly 
ingesting changes coming from different databases using Change Data 
Capture(CDC).
+
+
+
+
+  What is 
Flink CDC?
+  
+  
+  
+Flink CDC is a distributed data integration tool for real time 
data and batch data. 
+Flink CDC brings the simplicity and elegance of data integration 
via YAML to describe
+the data movement and transformation.
+  
+  
+  {{< img src="/fig/index-yaml-example.png" alt="Flink CDC Example">}}
+  
+
+
+
+ 
 
 
-Flink CDC integrates Debezium as the engine to capture data changes. So it can 
fully leverage the ability of Debezium. See more about what is 
[Debezium](https://github.com/debezium/debezium).
 
-{{< img src="/fig/cdc-flow.png" alt="Stateful Functions" width="50%" >}}
+
+
+
+
+
+  Key 
Features
+
+
+
+
+
+
+
+
+
+
+  Change Data Capture
+  
+  
+FLink CDC supports distributed scanning of historical data of 
database and then automatically switches to change data capturing. The switch 
uses the incremental snapshot algorithm which ensure the switch action does not 
lock the database.  
+  
+
+
+
+  Schema Evolution
+  
+  
+Flink CDC will automatically create downstream table using the 
inferred table structure based on upstream table. Flink CDC will automatically 
apply upstream DDL to downstream systems during change data capturing.
+
+
+
+  Streaming Pipeline
+  
+  
+Flink CDC jobs run in streaming mode by default, providing 
sub-second end-to-end latency in real-time binlog synchronization scenarios, 
this feature effectively ensures data freshness for downstream businesses.
+
+
+
+
+
+
+
+
+
+
+
+
+  Data Transformation
+  
+  
+Flink CDC will soon support data transform operations of ETL, 
including column projection, computed column, filter expression and classical 
scalar functions.
+
+
+
+  Full Database Sync
+  
+  Flink CDC supports synchronizing all 
tables of source database instance to downstream in one job, user can config 
the captured database list and table list.

Review Comment:
   ```suggestion
 Flink CDC supports synchronizing all 
tables of source database instance to downstream in one job by configuring the 
captured database list and table list.
   ```



##
docs/content/_index.md:
##
@@ -22,17 +22,179 @@ specific language governing permissions and limitations
 under the License.
 -->
 
- 
+
+
+
+
+Flink CDC
+
+
+The streaming data integration tool
+
+
+
+
+
+
+
+
+
+
+
+  
+  {{< img src="/fig/cdc-flow.png" alt="Flink CDC Flow">}}
+  
+
+
+
+
+
 
-
-  
-Flink CDC: Change Data Capture Solution Of Apache Flink
-  
-Set of source connectors for Apache Flink® directly 
ingesting changes coming from different databases using Change Data 
Capture(CDC).
+
+
+
+
+  What is 
Flink CDC?
+  
+  
+  
+Flink CDC is a distributed data integration tool for real time 
data and batch data. 
+Flink CDC brings the simplicity and elegance of data integration 
via YAML to describe
+the data movement and transformation.
+  
+  
+  {{< img src="/fig/index-yaml-example.png" alt="Flink CDC Example">}}
+  
+
+
+
+ 
 
 
-Flink CDC integrates Debezium as the engine to capture data changes. So it can 
fully leverage the ability of Debezium. See more about what is 
[Debezium](https://github.com/debezium/debezium).
 
-{{< img src="/fig/cdc-flow.png" alt="Stateful Functions" width="50%" >}}
+
+
+
+
+
+  Key 
Features
+
+
+
+
+
+
+
+
+
+
+  Change Data Capture
+  
+  
+FLink CDC supports distributed scanning of historical data of 
database and then automatically switches to change data capturing. The s

[PR] [FLINK-34684 & 34683][docs] Add licenses & contributing docs in Flink CDC [flink-cdc]

2024-03-17 Thread via GitHub


ruanhang1993 opened a new pull request, #3159:
URL: https://github.com/apache/flink-cdc/pull/3159

   This PR adds licenses & contributing docs to Flink CDC.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34684) "Developer Guide - Licenses" Page for Flink CDC Documentation

2024-03-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34684?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34684:
---
Labels: pull-request-available  (was: )

> "Developer Guide - Licenses" Page for Flink CDC Documentation
> -
>
> Key: FLINK-34684
> URL: https://issues.apache.org/jira/browse/FLINK-34684
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Qingsheng Ren
>Assignee: Hang Ruan
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.1.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [cdc][cli][2934]Submit CDC Job To Flink K8S Native Application Mode [flink-cdc]

2024-03-17 Thread via GitHub


czy006 commented on code in PR #3093:
URL: https://github.com/apache/flink-cdc/pull/3093#discussion_r1527747030


##
flink-cdc-cli/src/main/java/org/apache/flink/cdc/cli/CliFrontendOptions.java:
##
@@ -46,6 +46,16 @@ public class CliFrontendOptions {
 .desc("JARs to be submitted together with the pipeline")
 .build();
 
+public static final Option TARGET =
+Option.builder("t")
+.longOpt("target")
+.hasArg()
+.desc(
+"The deployment target for the execution. This can 
take one of the following values "

Review Comment:
   https://issues.apache.org/jira/browse/FLINK-34677 Maybe can in there, WDHY?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34648] add waitChangeResultRequest and WaitChangeResultResponse to avoid RPC timeout. [flink-cdc]

2024-03-17 Thread via GitHub


fanqiejiang8 commented on PR #3128:
URL: https://github.com/apache/flink-cdc/pull/3128#issuecomment-2002855730

   @lvyanquan @BaoPiao When task parallelism is multiple, task retry all sent 
SchemaRegistry task more the RefreshPendingListsRequest request, that each 
request will not cause data inconsistency pendingSchemaChanges
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [build] fix jackson conflicts among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


link3280 commented on PR #2987:
URL: https://github.com/apache/flink-cdc/pull/2987#issuecomment-2002817974

   ping @leonardBang 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [build] fix inconsistent Kafka shading among cdc connectors [flink-cdc]

2024-03-17 Thread via GitHub


link3280 commented on PR #2988:
URL: https://github.com/apache/flink-cdc/pull/2988#issuecomment-2002816885

   ping @leonardBang 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34679][cdc] add doc under core-concept. [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on PR #3153:
URL: https://github.com/apache/flink-cdc/pull/3153#issuecomment-2002777037

   @leonardBang  @PatrickRen PTAL.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-34678) "Introduction" Page for Flink CDC Documentation

2024-03-17 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34678?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-34678:
---
Labels: pull-request-available  (was: )

> "Introduction" Page for Flink CDC Documentation
> ---
>
> Key: FLINK-34678
> URL: https://issues.apache.org/jira/browse/FLINK-34678
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Flink CDC
>Affects Versions: cdc-3.1.0
>Reporter: Qingsheng Ren
>Assignee: Qingsheng Ren
>Priority: Major
>  Labels: pull-request-available
> Fix For: cdc-3.1.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] [FLINK-34678][cdc] Add introduction page for Flink CDC docs [flink-cdc]

2024-03-17 Thread via GitHub


PatrickRen opened a new pull request, #3158:
URL: https://github.com/apache/flink-cdc/pull/3158

   This pull request adds introduction page for Flink CDC docs.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32074][checkpoint] Merge file across checkpoints [flink]

2024-03-17 Thread via GitHub


masteryhx commented on PR #24497:
URL: https://github.com/apache/flink/pull/24497#issuecomment-2002763228

   Thanks for the update.
   Merged 841f23c7 into master.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-32074) Support file merging across checkpoints

2024-03-17 Thread Hangxiang Yu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32074?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hangxiang Yu resolved FLINK-32074.
--
Resolution: Fixed

Merged 841f23c7 into master

> Support file merging across checkpoints
> ---
>
> Key: FLINK-32074
> URL: https://issues.apache.org/jira/browse/FLINK-32074
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / Checkpointing, Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Zakelly Lan
>Assignee: Zakelly Lan
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.20.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32074][checkpoint] Merge file across checkpoints [flink]

2024-03-17 Thread via GitHub


masteryhx closed pull request #24497: [FLINK-32074][checkpoint] Merge file 
across checkpoints
URL: https://github.com/apache/flink/pull/24497


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Resolved] (FLINK-34334) Add sub-task level RocksDB file count metric

2024-03-17 Thread Hangxiang Yu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34334?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hangxiang Yu resolved FLINK-34334.
--
Resolution: Fixed

Merged 
[{{0921968}}|https://github.com/apache/flink/commit/0921968bb7d38bfac0b7899ec974a9744a721b22]
 into master.

> Add sub-task level RocksDB file count metric
> 
>
> Key: FLINK-34334
> URL: https://issues.apache.org/jira/browse/FLINK-34334
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Jufang He
>Assignee: Jufang He
>Priority: Major
>  Labels: pull-request-available
> Attachments: img_v3_027i_7ed0b8ba-3f12-48eb-aab3-cc368ac47cdg.jpg
>
>
> In our production environment, we encountered the problem of task deploy 
> failure. The root cause was that too many sst files of a single sub-task led 
> to too much task deployment information(OperatorSubtaskState), and then 
> caused akka request timeout in the task deploy phase. Therefore, I wanted to 
> add sub-task level RocksDB file count metrics. It is convenient to avoid 
> performance problems caused by too many sst files in time.
> RocksDB has provided the JNI 
> (https://javadoc.io/doc/org.rocksdb/rocksdbjni/6.20.3/org/rocksdb/RocksDB.html#getColumnFamilyMetaData
>  ()) We can easily retrieve the file count and report it via metrics reporter.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34334][state] Add sub-task level RocksDB file count metrics [flink]

2024-03-17 Thread via GitHub


masteryhx closed pull request #24322: [FLINK-34334][state] Add sub-task level 
RocksDB file count metrics
URL: https://github.com/apache/flink/pull/24322


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Add announcement blog post for Flink 1.19 [flink-web]

2024-03-17 Thread via GitHub


lincoln-lil commented on PR #721:
URL: https://github.com/apache/flink-web/pull/721#issuecomment-2002757799

   Finalize the commits before merging...


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34687][cdc][docs] Build the home page of Flink CDC documentation [flink-cdc]

2024-03-17 Thread via GitHub


xleoken commented on PR #3156:
URL: https://github.com/apache/flink-cdc/pull/3156#issuecomment-2002754671

   👍 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34671][cdc] update README.md file to update links and description. [flink-cdc]

2024-03-17 Thread via GitHub


LYanquan commented on PR #3152:
URL: https://github.com/apache/flink-cdc/pull/3152#issuecomment-2002750194

   @leonardBang @PatrickRen  PTAL.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-20578) Cannot create empty array using ARRAY[]

2024-03-17 Thread Nathan Taylor Armstrong Lewis (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17827826#comment-17827826
 ] 

Nathan Taylor Armstrong Lewis commented on FLINK-20578:
---

Does anyone know of a workaround to create an empty array literal until this 
issue is addressed?

> Cannot create empty array using ARRAY[]
> ---
>
> Key: FLINK-20578
> URL: https://issues.apache.org/jira/browse/FLINK-20578
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / API
>Affects Versions: 1.11.2
>Reporter: Fabian Hueske
>Assignee: Eric Xiao
>Priority: Major
>  Labels: pull-request-available, stale-assigned, starter
> Fix For: 1.20.0
>
> Attachments: Screen Shot 2022-10-25 at 10.50.42 PM.png, Screen Shot 
> 2022-10-25 at 10.50.47 PM.png, Screen Shot 2022-10-25 at 11.01.06 PM.png, 
> Screen Shot 2022-10-26 at 2.28.49 PM.png, image-2022-10-26-14-42-08-468.png, 
> image-2022-10-26-14-42-57-579.png
>
>
> Calling the ARRAY function without an element (`ARRAY[]`) results in an error 
> message.
> Is that the expected behavior?
> How can users create empty arrays?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34660][checkpoint] Parse cluster configuration for AutoRescalingITCase#testCheckpointRescalingInKeyedState [flink]

2024-03-17 Thread via GitHub


masteryhx commented on PR #24504:
URL: https://github.com/apache/flink/pull/24504#issuecomment-2002742851

   Thanks for the review.
   Merged e79df354 into master.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-34660][checkpoint] Parse cluster configuration for AutoRescalingITCase#testCheckpointRescalingInKeyedState [flink]

2024-03-17 Thread via GitHub


masteryhx closed pull request #24504: [FLINK-34660][checkpoint] Parse cluster 
configuration for AutoRescalingITCase#testCheckpointRescalingInKeyedState
URL: https://github.com/apache/flink/pull/24504


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-34571) SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError failed due an assertion

2024-03-17 Thread Weijie Guo (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34571?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Weijie Guo closed FLINK-34571.
--
Fix Version/s: 1.19.1
   Resolution: Fixed

> SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError failed 
> due an assertion
> --
>
> Key: FLINK-34571
> URL: https://issues.apache.org/jira/browse/FLINK-34571
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.19.0, 1.20.0
>Reporter: Matthias Pohl
>Assignee: Weijie Guo
>Priority: Critical
>  Labels: github-actions, pull-request-available, test-stability
> Fix For: 1.20.0, 1.19.1
>
>
> https://github.com/apache/flink/actions/runs/8134965216/job/8875618#step:10:8586
> {code}
> Error: 02:39:36 02:39:36.688 [ERROR] Tests run: 9, Failures: 1, Errors: 0, 
> Skipped: 0, Time elapsed: 13.68 s <<< FAILURE! -- in 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest
> Error: 02:39:36 02:39:36.689 [ERROR] 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError
>  -- Time elapsed: 0.174 s <<< FAILURE!
> Mar 04 02:39:36 org.opentest4j.AssertionFailedError: 
> Mar 04 02:39:36 
> Mar 04 02:39:36 Expecting value to be true but was false
> Mar 04 02:39:36   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> Mar 04 02:39:36   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> Mar 04 02:39:36   at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> Mar 04 02:39:36   at 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError(SortMergeResultPartitionReadSchedulerTest.java:225)
> Mar 04 02:39:36   at java.lang.reflect.Method.invoke(Method.java:498)
> Mar 04 02:39:36   at 
> java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-34571) SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError failed due an assertion

2024-03-17 Thread Weijie Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-34571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17825198#comment-17825198
 ] 

Weijie Guo edited comment on FLINK-34571 at 3/18/24 1:36 AM:
-

master(1.20) via fcf61b9108942a892aa93765178a81f2c0e787c5.
release-1.19: 4f7f6a97f799a4a30d1aea57fc0af5d07dc0eeed.


was (Author: weijie guo):
master(1.20) via fcf61b9108942a892aa93765178a81f2c0e787c5.

Will backport this to release-1.19 after it released.

> SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError failed 
> due an assertion
> --
>
> Key: FLINK-34571
> URL: https://issues.apache.org/jira/browse/FLINK-34571
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Network
>Affects Versions: 1.19.0, 1.20.0
>Reporter: Matthias Pohl
>Assignee: Weijie Guo
>Priority: Critical
>  Labels: github-actions, pull-request-available, test-stability
> Fix For: 1.20.0
>
>
> https://github.com/apache/flink/actions/runs/8134965216/job/8875618#step:10:8586
> {code}
> Error: 02:39:36 02:39:36.688 [ERROR] Tests run: 9, Failures: 1, Errors: 0, 
> Skipped: 0, Time elapsed: 13.68 s <<< FAILURE! -- in 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest
> Error: 02:39:36 02:39:36.689 [ERROR] 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError
>  -- Time elapsed: 0.174 s <<< FAILURE!
> Mar 04 02:39:36 org.opentest4j.AssertionFailedError: 
> Mar 04 02:39:36 
> Mar 04 02:39:36 Expecting value to be true but was false
> Mar 04 02:39:36   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> Mar 04 02:39:36   at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> Mar 04 02:39:36   at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> Mar 04 02:39:36   at 
> org.apache.flink.runtime.io.network.partition.SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError(SortMergeResultPartitionReadSchedulerTest.java:225)
> Mar 04 02:39:36   at java.lang.reflect.Method.invoke(Method.java:498)
> Mar 04 02:39:36   at 
> java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
> Mar 04 02:39:36   at 
> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [BP-1.19][FLINK-34571][test] Fix flaky test SortMergeResultPartitionReadSchedulerTest.testOnReadBufferRequestError [flink]

2024-03-17 Thread via GitHub


reswqa merged PR #24479:
URL: https://github.com/apache/flink/pull/24479


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [Flink 34569][e2e] fail fast if AWS cli container fails to start [flink]

2024-03-17 Thread via GitHub


robobario commented on code in PR #24491:
URL: https://github.com/apache/flink/pull/24491#discussion_r1527607029


##
flink-end-to-end-tests/test-scripts/common_s3_operations.sh:
##
@@ -58,7 +64,11 @@ function aws_cli_stop() {
 if [[ $AWSCLI_CONTAINER_ID ]]; then
   aws_cli_stop
 fi
-aws_cli_start
+aws_cli_start || aws_cli_start

Review Comment:
   I've added in a failsafe to kill/remove if `CONTAINER_ID` is non-empty after 
the docker run fails.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-34633) Support unnesting array constants

2024-03-17 Thread Xingcan Cui (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-34633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xingcan Cui reassigned FLINK-34633:
---

Assignee: Jeyhun Karimov

> Support unnesting array constants
> -
>
> Key: FLINK-34633
> URL: https://issues.apache.org/jira/browse/FLINK-34633
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Planner
>Affects Versions: 1.18.1
>Reporter: Xingcan Cui
>Assignee: Jeyhun Karimov
>Priority: Minor
>  Labels: pull-request-available
>
> It seems that the current planner doesn't support using UNNEST on array 
> constants.(x)
> {code:java}
> SELECT * FROM UNNEST(ARRAY[1,2,3]);{code}
>  
> The following query can't be compiled.(x)
> {code:java}
> SELECT * FROM (VALUES('a')) CROSS JOIN UNNEST(ARRAY[1, 2, 3]){code}
>  
> The rewritten version works. (/)
> {code:java}
> SELECT * FROM (SELECT *, ARRAY[1,2,3] AS A FROM (VALUES('a'))) CROSS JOIN 
> UNNEST(A){code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-34633][table] Support unnesting array constants [flink]

2024-03-17 Thread via GitHub


flinkbot commented on PR #24510:
URL: https://github.com/apache/flink/pull/24510#issuecomment-2002539941

   
   ## CI report:
   
   * e239874a6d95df04d4c5919b09fc800b82cacc44 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



  1   2   >