[flink-ml] branch master updated: [FLINK-31255] OperatorUtils#createWrappedOperatorConfig should update input and sideOutput serializers

2023-04-17 Thread lindong
This is an automated email from the ASF dual-hosted git repository.

lindong pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink-ml.git


The following commit(s) were added to refs/heads/master by this push:
 new 89468545 [FLINK-31255] OperatorUtils#createWrappedOperatorConfig 
should update input and sideOutput serializers
89468545 is described below

commit 894685455d1c26fd45198857b7a96ee850725a59
Author: JiangXin 
AuthorDate: Tue Apr 18 13:47:52 2023 +0800

[FLINK-31255] OperatorUtils#createWrappedOperatorConfig should update input 
and sideOutput serializers

This closes #229.
---
 .../flink/iteration/operator/OperatorUtils.java| 65 +++--
 flink-ml-tests/pom.xml |  7 +++
 .../iteration/UnboundedStreamIterationITCase.java  | 67 ++
 3 files changed, 135 insertions(+), 4 deletions(-)

diff --git 
a/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java
 
b/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java
index b3b629fe..3d67b7fa 100644
--- 
a/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java
+++ 
b/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java
@@ -18,21 +18,26 @@
 
 package org.apache.flink.iteration.operator;
 
+import org.apache.flink.api.common.typeutils.TypeSerializer;
 import org.apache.flink.api.java.functions.KeySelector;
 import org.apache.flink.configuration.Configuration;
 import org.apache.flink.core.fs.Path;
 import org.apache.flink.iteration.IterationID;
 import org.apache.flink.iteration.config.IterationOptions;
 import org.apache.flink.iteration.proxy.ProxyKeySelector;
+import org.apache.flink.iteration.typeinfo.IterationRecordSerializer;
+import org.apache.flink.iteration.typeinfo.IterationRecordTypeInfo;
 import org.apache.flink.iteration.utils.ReflectionUtils;
 import org.apache.flink.runtime.jobgraph.OperatorID;
 import org.apache.flink.statefun.flink.core.feedback.FeedbackChannel;
 import org.apache.flink.statefun.flink.core.feedback.FeedbackConsumer;
 import org.apache.flink.statefun.flink.core.feedback.FeedbackKey;
 import org.apache.flink.streaming.api.graph.StreamConfig;
+import org.apache.flink.streaming.api.graph.StreamConfig.NetworkInputConfig;
 import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator;
 import org.apache.flink.streaming.api.operators.StreamOperator;
 import org.apache.flink.util.ExceptionUtils;
+import org.apache.flink.util.OutputTag;
 import org.apache.flink.util.function.SupplierWithException;
 import org.apache.flink.util.function.ThrowingConsumer;
 
@@ -42,6 +47,7 @@ import java.util.Arrays;
 import java.util.Random;
 import java.util.UUID;
 import java.util.concurrent.Executor;
+import java.util.stream.Stream;
 
 import static org.apache.flink.util.Preconditions.checkState;
 
@@ -89,11 +95,10 @@ public class OperatorUtils {
 }
 }
 
-public static StreamConfig createWrappedOperatorConfig(
-StreamConfig wrapperConfig, ClassLoader cl) {
-StreamConfig wrappedConfig = new 
StreamConfig(wrapperConfig.getConfiguration().clone());
+public static StreamConfig createWrappedOperatorConfig(StreamConfig 
config, ClassLoader cl) {
+StreamConfig wrappedConfig = new 
StreamConfig(config.getConfiguration().clone());
 for (int i = 0; i < wrappedConfig.getNumberOfNetworkInputs(); ++i) {
-KeySelector keySelector = wrapperConfig.getStatePartitioner(i, cl);
+KeySelector keySelector = config.getStatePartitioner(i, cl);
 if (keySelector != null) {
 checkState(
 keySelector instanceof ProxyKeySelector,
@@ -104,6 +109,58 @@ public class OperatorUtils {
 }
 }
 
+StreamConfig.InputConfig[] inputs = config.getInputs(cl);
+for (int i = 0; i < inputs.length; ++i) {
+if (inputs[i] instanceof NetworkInputConfig) {
+TypeSerializer typeSerializerIn =
+((NetworkInputConfig) inputs[i]).getTypeSerializer();
+checkState(
+typeSerializerIn instanceof IterationRecordSerializer,
+"The serializer of input[%s] should be 
IterationRecordSerializer but it is %s.",
+i,
+typeSerializerIn);
+inputs[i] =
+new NetworkInputConfig(
+((IterationRecordSerializer) 
typeSerializerIn)
+.getInnerSerializer(),
+i);
+}
+}
+wrappedConfig.setInputs(inputs);
+
+TypeSerializer typeSerializerOut = config.getTypeSerializerOut(cl);
+checkState(
+typeSerializerOut instanceof IterationRecordSerializer,
+

[flink-connector-jdbc] 03/03: [FLINK-31649] Fix archunit connectors options must be public

2023-04-17 Thread wanglijie
This is an automated email from the ASF dual-hosted git repository.

wanglijie pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git

commit 396f205bcbed133d1e7dc0d6c3ce9487dcbc1dd7
Author: Joao Boto 
AuthorDate: Thu Apr 13 15:00:33 2023 +0200

[FLINK-31649] Fix archunit connectors options must be public
---
 .../b10c194a-0517-4744-830f-81abfb71ab18   |  2 --
 .../connector/jdbc/internal/JdbcOutputFormat.java  |  6 ++--
 ...ons.java => InternalJdbcConnectionOptions.java} | 14 
 .../jdbc/table/JdbcDynamicTableFactory.java| 13 
 .../connector/jdbc/table/JdbcDynamicTableSink.java |  6 ++--
 .../jdbc/table/JdbcDynamicTableSource.java |  6 ++--
 .../jdbc/table/JdbcOutputFormatBuilder.java|  6 ++--
 .../jdbc/table/JdbcRowDataLookupFunction.java  |  4 +--
 .../connector/jdbc/internal/JdbcFullTest.java  | 10 +++---
 .../jdbc/internal/JdbcTableOutputFormatTest.java   | 12 +++
 .../jdbc/table/JdbcAppendOnlyWriterTest.java   |  4 +--
 .../jdbc/table/JdbcDynamicTableFactoryTest.java| 30 -
 .../connector/jdbc/table/JdbcOutputFormatTest.java | 38 +++---
 .../jdbc/table/JdbcRowDataInputFormatTest.java |  4 +--
 .../jdbc/table/JdbcRowDataLookupFunctionTest.java  |  6 ++--
 15 files changed, 80 insertions(+), 81 deletions(-)

diff --git 
a/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18 
b/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18
index 400c942..e69de29 100644
--- 
a/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18
+++ 
b/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18
@@ -1,2 +0,0 @@
-Class  
does not reside in a package 'org.apache.flink..table' in 
(JdbcConnectorOptions.java:0)
-org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions does not 
satisfy: annotated with @PublicEvolving or annotated with @Public
\ No newline at end of file
diff --git 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java
 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java
index a869f6c..1ca8b63 100644
--- 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java
+++ 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java
@@ -33,7 +33,7 @@ import org.apache.flink.connector.jdbc.JdbcStatementBuilder;
 import 
org.apache.flink.connector.jdbc.internal.connection.JdbcConnectionProvider;
 import 
org.apache.flink.connector.jdbc.internal.connection.SimpleJdbcConnectionProvider;
 import 
org.apache.flink.connector.jdbc.internal.executor.JdbcBatchStatementExecutor;
-import org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions;
+import 
org.apache.flink.connector.jdbc.internal.options.InternalJdbcConnectionOptions;
 import org.apache.flink.connector.jdbc.internal.options.JdbcDmlOptions;
 import 
org.apache.flink.connector.jdbc.statement.FieldNamedPreparedStatementImpl;
 import org.apache.flink.connector.jdbc.utils.JdbcUtils;
@@ -284,7 +284,7 @@ public class JdbcOutputFormat new NullPointerException("No driverName 
supplied."));
 }
 
-return new JdbcConnectorOptions(
+return new InternalJdbcConnectionOptions(
 dialect.appendDefaultUrlProperties(dbURL),
 tableName,
 driverName,
diff --git 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java
 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java
index bff5a49..3434f60 100644
--- 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java
+++ 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java
@@ -25,7 +25,7 @@ import org.apache.flink.configuration.ReadableConfig;
 import org.apache.flink.connector.jdbc.JdbcExecutionOptions;
 import org.apache.flink.connector.jdbc.dialect.JdbcDialect;
 import org.apache.flink.connector.jdbc.dialect.JdbcDialectLoader;
-import org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions;
+import 
org.apache.flink.connector.jdbc.internal.options.InternalJdbcConnectionOptions;
 import org.apache.flink.connector.jdbc.internal.options.JdbcDmlOptions;
 import org.apache.flink.connector.jdbc.internal.options.JdbcReadOptions;
 import org.apache.flink.table.connector.sink.DynamicTableSink;
@@ -90,7 +90,8 @@ public class JdbcDynamicTableFactory implements 
DynamicTableSourceFactory, Dynam
 validateConfigOptions(config, context.getClassLoader());
 validateDataTypeWithJdbcDialect(
 context.getPhysicalRowDataType(), config.get(URL), 
context.getClassLoader());
-  

[flink-connector-jdbc] branch main updated (b6c3cb2 -> 396f205)

2023-04-17 Thread wanglijie
This is an automated email from the ASF dual-hosted git repository.

wanglijie pushed a change to branch main
in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git


from b6c3cb2  [FLINK-31759][JDBC] Using sql_connector_download_table 
shortcode for table document.
 new 508b152  [FLINK-31649] Fix archunit ITCase needs minicluster
 new 6397c63  [FLINK-31649] Fix archunit returning public(evolving) methods
 new 396f205  [FLINK-31649] Fix archunit connectors options must be public

The 3 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 .../6b9ab1b0-c14d-4667-bab5-407b81fba98b   | 36 
 .../b10c194a-0517-4744-830f-81abfb71ab18   |  2 --
 .../d45c3af5-52c6-45fd-9926-75e75e77473a   |  7 
 .../connector/jdbc/JdbcExactlyOnceOptions.java |  1 +
 .../flink/connector/jdbc/JdbcExecutionOptions.java |  1 +
 .../flink/connector/jdbc/catalog/JdbcCatalog.java  |  2 ++
 .../connector/jdbc/dialect/AbstractDialect.java|  1 +
 .../connector/jdbc/internal/JdbcOutputFormat.java  |  6 ++--
 ...ons.java => InternalJdbcConnectionOptions.java} | 14 
 .../jdbc/table/JdbcDynamicTableFactory.java| 13 
 .../connector/jdbc/table/JdbcDynamicTableSink.java |  6 ++--
 .../jdbc/table/JdbcDynamicTableSource.java |  6 ++--
 .../jdbc/table/JdbcOutputFormatBuilder.java|  6 ++--
 .../jdbc/table/JdbcRowDataLookupFunction.java  |  4 +--
 .../connector/jdbc/xa/JdbcXaSinkFunctionState.java |  3 ++
 .../connector/jdbc/xa/XaSinkStateHandler.java  |  4 +--
 .../apache/flink/connector/jdbc/JdbcITCase.java|  3 +-
 .../jdbc/catalog/MySqlCatalogTestBase.java |  3 +-
 .../jdbc/catalog/PostgresCatalogTestBase.java  |  3 +-
 .../connector/jdbc/internal/JdbcFullTest.java  | 10 +++---
 .../jdbc/internal/JdbcTableOutputFormatTest.java   | 12 +++
 ...ProviderDriverClassConcurrentLoadingITCase.java |  3 +-
 .../jdbc/table/JdbcAppendOnlyWriterTest.java   |  4 +--
 .../jdbc/table/JdbcDynamicTableFactoryTest.java| 30 -
 .../connector/jdbc/table/JdbcOutputFormatTest.java | 38 +++---
 .../jdbc/table/JdbcRowDataInputFormatTest.java |  4 +--
 .../jdbc/table/JdbcRowDataLookupFunctionTest.java  |  6 ++--
 .../connector/jdbc/testutils/JdbcITCaseBase.java}  | 30 -
 28 files changed, 111 insertions(+), 147 deletions(-)
 rename 
flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/options/{JdbcConnectorOptions.java
 => InternalJdbcConnectionOptions.java} (94%)
 copy 
flink-connector-jdbc/src/{main/java/org/apache/flink/connector/jdbc/internal/converter/DerbyRowConverter.java
 => test/java/org/apache/flink/connector/jdbc/testutils/JdbcITCaseBase.java} 
(55%)



[flink-connector-jdbc] 02/03: [FLINK-31649] Fix archunit returning public(evolving) methods

2023-04-17 Thread wanglijie
This is an automated email from the ASF dual-hosted git repository.

wanglijie pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git

commit 6397c637e8d2b5c5223b769b453692b8f03ae927
Author: Joao Boto 
AuthorDate: Mon Apr 17 11:35:23 2023 +0200

[FLINK-31649] Fix archunit returning public(evolving) methods
---
 .../archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a   | 7 ---
 .../org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java| 1 +
 .../java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java | 1 +
 .../java/org/apache/flink/connector/jdbc/catalog/JdbcCatalog.java  | 2 ++
 .../org/apache/flink/connector/jdbc/dialect/AbstractDialect.java   | 1 +
 .../apache/flink/connector/jdbc/xa/JdbcXaSinkFunctionState.java| 3 +++
 .../org/apache/flink/connector/jdbc/xa/XaSinkStateHandler.java | 4 ++--
 7 files changed, 10 insertions(+), 9 deletions(-)

diff --git 
a/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a 
b/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a
index 4f453d5..e69de29 100644
--- 
a/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a
+++ 
b/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a
@@ -1,7 +0,0 @@
-org.apache.flink.connector.jdbc.JdbcExactlyOnceOptions.builder(): Returned 
leaf type 
org.apache.flink.connector.jdbc.JdbcExactlyOnceOptions$JDBCExactlyOnceOptionsBuilder
 does not satisfy: reside outside of package 'org.apache.flink..' or reside in 
any package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
-org.apache.flink.connector.jdbc.JdbcExecutionOptions.builder(): Returned leaf 
type org.apache.flink.connector.jdbc.JdbcExecutionOptions$Builder does not 
satisfy: reside outside of package 'org.apache.flink..' or reside in any 
package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
-org.apache.flink.connector.jdbc.catalog.JdbcCatalog.getInternal(): Returned 
leaf type org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog does not 
satisfy: reside outside of package 'org.apache.flink..' or reside in any 
package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
-org.apache.flink.connector.jdbc.dialect.AbstractDialect.decimalPrecisionRange():
 Returned leaf type 
org.apache.flink.connector.jdbc.dialect.AbstractDialect$Range does not satisfy: 
reside outside of package 'org.apache.flink..' or reside in any package 
['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or 
annotated with @Deprecated
-org.apache.flink.connector.jdbc.dialect.AbstractDialect.timestampPrecisionRange():
 Returned leaf type 
org.apache.flink.connector.jdbc.dialect.AbstractDialect$Range does not satisfy: 
reside outside of package 'org.apache.flink..' or reside in any package 
['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or 
annotated with @Deprecated
-org.apache.flink.connector.jdbc.xa.XaSinkStateHandler.load(org.apache.flink.runtime.state.FunctionInitializationContext):
 Returned leaf type org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState 
does not satisfy: reside outside of package 'org.apache.flink..' or reside in 
any package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
-org.apache.flink.connector.jdbc.xa.XaSinkStateHandler.store(org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState):
 Argument leaf type org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState 
does not satisfy: reside outside of package 'org.apache.flink..' or reside in 
any package ['..shaded..'] or annotated with @Public or annotated with 
@PublicEvolving or annotated with @Deprecated
\ No newline at end of file
diff --git 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java
 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java
index c17ed17..47332d4 100644
--- 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java
+++ 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java
@@ -104,6 +104,7 @@ public class JdbcExactlyOnceOptions implements Serializable 
{
 }
 
 /** JDBCExactlyOnceOptionsBuilder. */
+@PublicEvolving
 public static class JDBCExactlyOnceOptionsBuilder {
 private boolean recoveredAndRollback = DEFAULT_RECOVERED_AND_ROLLBACK;
 private int maxCommitAttempts = DEFAULT_MAX_COMMIT_ATTEMPTS;
diff --git 
a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java
 
b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java
index 137fb7c..c19677b 100644
--- 

[flink-connector-jdbc] 01/03: [FLINK-31649] Fix archunit ITCase needs minicluster

2023-04-17 Thread wanglijie
This is an automated email from the ASF dual-hosted git repository.

wanglijie pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git

commit 508b152bf7ee50643b6b3844f20289b6461e3c25
Author: Joao Boto 
AuthorDate: Tue Mar 28 17:18:35 2023 +0200

[FLINK-31649] Fix archunit ITCase needs minicluster
---
 .../6b9ab1b0-c14d-4667-bab5-407b81fba98b   | 36 --
 .../apache/flink/connector/jdbc/JdbcITCase.java|  3 +-
 .../jdbc/catalog/MySqlCatalogTestBase.java |  3 +-
 .../jdbc/catalog/PostgresCatalogTestBase.java  |  3 +-
 ...ProviderDriverClassConcurrentLoadingITCase.java |  3 +-
 .../connector/jdbc/testutils/JdbcITCaseBase.java   | 36 ++
 6 files changed, 44 insertions(+), 40 deletions(-)

diff --git 
a/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b 
b/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b
index e54d3ff..e69de29 100644
--- 
a/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b
+++ 
b/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b
@@ -1,36 +0,0 @@
-org.apache.flink.connector.jdbc.JdbcITCase does not satisfy: only one of the 
following predicates match:\
-* reside in a package 'org.apache.flink.runtime.*' and contain any fields that 
are static, final, and of type InternalMiniClusterExtension and annotated with 
@RegisterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and contain any 
fields that are static, final, and of type MiniClusterExtension and annotated 
with @RegisterExtension\
-* reside in a package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class InternalMiniClusterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class MiniClusterExtension\
- or contain any fields that are public, static, and of type 
MiniClusterWithClientResource and final and annotated with @ClassRule or 
contain any fields that is of type MiniClusterWithClientResource and public and 
final and not static and annotated with @Rule
-org.apache.flink.connector.jdbc.catalog.MySqlCatalogITCase does not satisfy: 
only one of the following predicates match:\
-* reside in a package 'org.apache.flink.runtime.*' and contain any fields that 
are static, final, and of type InternalMiniClusterExtension and annotated with 
@RegisterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and contain any 
fields that are static, final, and of type MiniClusterExtension and annotated 
with @RegisterExtension\
-* reside in a package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class InternalMiniClusterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class MiniClusterExtension\
- or contain any fields that are public, static, and of type 
MiniClusterWithClientResource and final and annotated with @ClassRule or 
contain any fields that is of type MiniClusterWithClientResource and public and 
final and not static and annotated with @Rule
-org.apache.flink.connector.jdbc.catalog.MySql57CatalogITCase does not satisfy: 
only one of the following predicates match:\
-* reside in a package 'org.apache.flink.runtime.*' and contain any fields that 
are static, final, and of type InternalMiniClusterExtension and annotated with 
@RegisterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and contain any 
fields that are static, final, and of type MiniClusterExtension and annotated 
with @RegisterExtension\
-* reside in a package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class InternalMiniClusterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class MiniClusterExtension\
- or contain any fields that are public, static, and of type 
MiniClusterWithClientResource and final and annotated with @ClassRule or 
contain any fields that is of type MiniClusterWithClientResource and public and 
final and not static and annotated with @Rule
-org.apache.flink.connector.jdbc.catalog.MySql56CatalogITCase does not satisfy: 
only one of the following predicates match:\
-* reside in a package 'org.apache.flink.runtime.*' and contain any fields that 
are static, final, and of type InternalMiniClusterExtension and annotated with 
@RegisterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and contain any 
fields that are static, final, and of type MiniClusterExtension and annotated 
with @RegisterExtension\
-* reside in a package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class InternalMiniClusterExtension\
-* reside outside of package 'org.apache.flink.runtime.*' and is annotated with 
@ExtendWith with class MiniClusterExtension\
- or contain any fields that are public, static, and of type 

[flink] branch release-1.17 updated: [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions

2023-04-17 Thread guoweijie
This is an automated email from the ASF dual-hosted git repository.

guoweijie pushed a commit to branch release-1.17
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/release-1.17 by this push:
 new fa2263d1de1 [FLINK-31818][docs] Fix incorrect description of 
security.kerberos.access.hadoopFileSystems in SecurityOptions
fa2263d1de1 is described below

commit fa2263d1de1789a5200f909333cfe2dc30104b2f
Author: L 
AuthorDate: Tue Apr 18 00:02:53 2023 +0900

[FLINK-31818][docs] Fix incorrect description of 
security.kerberos.access.hadoopFileSystems in SecurityOptions
---
 docs/layouts/shortcodes/generated/security_auth_kerberos_section.html | 2 +-
 docs/layouts/shortcodes/generated/security_configuration.html | 2 +-
 .../src/main/java/org/apache/flink/configuration/SecurityOptions.java | 4 ++--
 3 files changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html 
b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
index ad063064b49..1e4e249148f 100644
--- a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
+++ b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
@@ -12,7 +12,7 @@
 security.kerberos.access.hadoopFileSystems
 (none)
 ListString
-A comma-separated list of Kerberos-secured Hadoop filesystems 
Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
+A semicolon-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
 
 
 security.kerberos.login.contexts
diff --git a/docs/layouts/shortcodes/generated/security_configuration.html 
b/docs/layouts/shortcodes/generated/security_configuration.html
index 5bfcc1710b3..069825bdd65 100644
--- a/docs/layouts/shortcodes/generated/security_configuration.html
+++ b/docs/layouts/shortcodes/generated/security_configuration.html
@@ -42,7 +42,7 @@
 security.kerberos.access.hadoopFileSystems
 (none)
 ListString
-A comma-separated list of Kerberos-secured Hadoop filesystems 
Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
+A semicolon-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
 
 
 security.kerberos.krb5-conf.path
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java 
b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
index 946894ad3f4..c9f0982e48f 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
@@ -167,8 +167,8 @@ public class SecurityOptions {
 .noDefaultValue()
 
.withDeprecatedKeys("yarn.security.kerberos.additionalFileSystems")
 .withDescription(
-"A comma-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, "
-+ 
"security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 "
+"A semicolon-separated list of Kerberos-secured 
Hadoop filesystems Flink is going to access. For example, "
++ 
"security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 "
 + "The JobManager needs to have access to 
these filesystems to retrieve the security tokens.");
 
 // 



[flink] branch master updated: [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions

2023-04-17 Thread guoweijie
This is an automated email from the ASF dual-hosted git repository.

guoweijie pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/master by this push:
 new fe8ff46123e [FLINK-31818][docs] Fix incorrect description of 
security.kerberos.access.hadoopFileSystems in SecurityOptions
fe8ff46123e is described below

commit fe8ff46123ee2f9278066c801228416e38e2cbd4
Author: L 
AuthorDate: Tue Apr 18 00:02:53 2023 +0900

[FLINK-31818][docs] Fix incorrect description of 
security.kerberos.access.hadoopFileSystems in SecurityOptions
---
 docs/layouts/shortcodes/generated/security_auth_kerberos_section.html | 2 +-
 docs/layouts/shortcodes/generated/security_configuration.html | 2 +-
 .../src/main/java/org/apache/flink/configuration/SecurityOptions.java | 4 ++--
 3 files changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html 
b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
index ad063064b49..1e4e249148f 100644
--- a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
+++ b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html
@@ -12,7 +12,7 @@
 security.kerberos.access.hadoopFileSystems
 (none)
 ListString
-A comma-separated list of Kerberos-secured Hadoop filesystems 
Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
+A semicolon-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
 
 
 security.kerberos.login.contexts
diff --git a/docs/layouts/shortcodes/generated/security_configuration.html 
b/docs/layouts/shortcodes/generated/security_configuration.html
index 5bfcc1710b3..069825bdd65 100644
--- a/docs/layouts/shortcodes/generated/security_configuration.html
+++ b/docs/layouts/shortcodes/generated/security_configuration.html
@@ -42,7 +42,7 @@
 security.kerberos.access.hadoopFileSystems
 (none)
 ListString
-A comma-separated list of Kerberos-secured Hadoop filesystems 
Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
+A semicolon-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, 
security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 The JobManager needs to have access to these filesystems to retrieve the 
security tokens.
 
 
 security.kerberos.krb5-conf.path
diff --git 
a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java 
b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
index 946894ad3f4..c9f0982e48f 100644
--- 
a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
+++ 
b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java
@@ -167,8 +167,8 @@ public class SecurityOptions {
 .noDefaultValue()
 
.withDeprecatedKeys("yarn.security.kerberos.additionalFileSystems")
 .withDescription(
-"A comma-separated list of Kerberos-secured Hadoop 
filesystems Flink is going to access. For example, "
-+ 
"security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003.
 "
+"A semicolon-separated list of Kerberos-secured 
Hadoop filesystems Flink is going to access. For example, "
++ 
"security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003.
 "
 + "The JobManager needs to have access to 
these filesystems to retrieve the security tokens.");
 
 // 



[flink] branch master updated: [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased

2023-04-17 Thread guoweijie
This is an automated email from the ASF dual-hosted git repository.

guoweijie pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/master by this push:
 new ffc6f3bfabd [FLINK-31763][runtime] Convert requested buffers to 
overdraft buffers when pool size is decreased
ffc6f3bfabd is described below

commit ffc6f3bfabd22b49b08f027400c194a8e7c9c51a
Author: Weijie Guo 
AuthorDate: Tue Apr 11 23:33:57 2023 +0800

[FLINK-31763][runtime] Convert requested buffers to overdraft buffers when 
pool size is decreased
---
 .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++-
 .../io/network/buffer/LocalBufferPoolTest.java | 85 --
 2 files changed, 59 insertions(+), 40 deletions(-)

diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
index 87b8a11e2aa..6506ab9f942 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
@@ -49,7 +49,11 @@ import static 
org.apache.flink.util.concurrent.FutureUtils.assertNoException;
  *
  * The size of this pool can be dynamically changed at runtime ({@link 
#setNumBuffers(int)}. It
  * will then lazily return the required number of buffers to the {@link 
NetworkBufferPool} to match
- * its new size.
+ * its new size. New buffers can be requested only when {@code 
numberOfRequestedMemorySegments +
+ * numberOfRequestedOverdraftMemorySegments < currentPoolSize + 
maxOverdraftBuffersPerGate}. In
+ * order to meet this requirement, when the size of this pool changes,
+ * numberOfRequestedMemorySegments and 
numberOfRequestedOverdraftMemorySegments can be converted to
+ * each other.
  *
  * Availability is defined as returning a non-overdraft segment on a 
subsequent {@link
  * #requestBuffer()}/ {@link #requestBufferBuilder()} and heaving a 
non-blocking {@link
@@ -671,13 +675,19 @@ class LocalBufferPool implements BufferPool {
 
 currentPoolSize = Math.min(numBuffers, maxNumberOfMemorySegments);
 
-// reset overdraft buffers
+// If pool size increases, try to convert overdraft buffer to 
ordinary buffer.
 while (numberOfRequestedOverdraftMemorySegments > 0
 && numberOfRequestedMemorySegments < currentPoolSize) {
 numberOfRequestedOverdraftMemorySegments--;
 numberOfRequestedMemorySegments++;
 }
 
+// If pool size decreases, try to convert ordinary buffer to 
overdraft buffer.
+while (numberOfRequestedMemorySegments > currentPoolSize) {
+numberOfRequestedMemorySegments--;
+numberOfRequestedOverdraftMemorySegments++;
+}
+
 returnExcessMemorySegments();
 
 if (isDestroyed) {
diff --git 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
index 956d55e5948..6c0fcf13b47 100644
--- 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
+++ 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
@@ -38,10 +38,8 @@ import java.util.ArrayDeque;
 import java.util.ArrayList;
 import java.util.Deque;
 import java.util.HashMap;
-import java.util.LinkedList;
 import java.util.List;
 import java.util.Map;
-import java.util.Queue;
 import java.util.concurrent.Callable;
 import java.util.concurrent.CompletableFuture;
 import java.util.concurrent.ExecutionException;
@@ -255,9 +253,38 @@ class LocalBufferPoolTest {
 void testDecreasePoolSize() throws Exception {
 final int maxMemorySegments = 10;
 final int requiredMemorySegments = 4;
-final int maxOverdraftBuffers = 2;
-final int largePoolSize = 5;
-final int smallPoolSize = 4;
+
+// requested buffers is equal to small pool size.
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 7, 5, 2, 5, 0, 5, 
0);
+// requested buffers is less than small pool size.
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 6, 4, 2, 2, 0, 3, 
1);
+// exceed buffers is equal to maxOverdraftBuffers
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 7, 5, 2, 7, 2, 5, 
0);
+// exceed buffers is greater than maxOverdraftBuffers
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 9, 5, 3, 9, 4, 5, 
0);
+// exceed buffers is less than maxOverdraftBuffers
+   

[flink] branch release-1.16 updated (09e8e57a100 -> ab2ba9612a6)

2023-04-17 Thread guoweijie
This is an automated email from the ASF dual-hosted git repository.

guoweijie pushed a change to branch release-1.16
in repository https://gitbox.apache.org/repos/asf/flink.git


from 09e8e57a100 [FLINK-31779][docs] Track stable branch of externalized 
connector instead of specific release tag
 add ab2ba9612a6 [FLINK-31763][runtime] Convert requested buffers to 
overdraft buffers when pool size is decreased

No new revisions were added by this update.

Summary of changes:
 .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++-
 .../io/network/buffer/LocalBufferPoolTest.java | 85 --
 2 files changed, 59 insertions(+), 40 deletions(-)



[flink] branch release-1.17 updated: [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased

2023-04-17 Thread guoweijie
This is an automated email from the ASF dual-hosted git repository.

guoweijie pushed a commit to branch release-1.17
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/release-1.17 by this push:
 new be0f9293c2c [FLINK-31763][runtime] Convert requested buffers to 
overdraft buffers when pool size is decreased
be0f9293c2c is described below

commit be0f9293c2ce00465154ef03f7cef29dd3116b8e
Author: Weijie Guo 
AuthorDate: Tue Apr 11 23:33:57 2023 +0800

[FLINK-31763][runtime] Convert requested buffers to overdraft buffers when 
pool size is decreased
---
 .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++-
 .../io/network/buffer/LocalBufferPoolTest.java | 85 --
 2 files changed, 59 insertions(+), 40 deletions(-)

diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
index 87b8a11e2aa..6506ab9f942 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java
@@ -49,7 +49,11 @@ import static 
org.apache.flink.util.concurrent.FutureUtils.assertNoException;
  *
  * The size of this pool can be dynamically changed at runtime ({@link 
#setNumBuffers(int)}. It
  * will then lazily return the required number of buffers to the {@link 
NetworkBufferPool} to match
- * its new size.
+ * its new size. New buffers can be requested only when {@code 
numberOfRequestedMemorySegments +
+ * numberOfRequestedOverdraftMemorySegments < currentPoolSize + 
maxOverdraftBuffersPerGate}. In
+ * order to meet this requirement, when the size of this pool changes,
+ * numberOfRequestedMemorySegments and 
numberOfRequestedOverdraftMemorySegments can be converted to
+ * each other.
  *
  * Availability is defined as returning a non-overdraft segment on a 
subsequent {@link
  * #requestBuffer()}/ {@link #requestBufferBuilder()} and heaving a 
non-blocking {@link
@@ -671,13 +675,19 @@ class LocalBufferPool implements BufferPool {
 
 currentPoolSize = Math.min(numBuffers, maxNumberOfMemorySegments);
 
-// reset overdraft buffers
+// If pool size increases, try to convert overdraft buffer to 
ordinary buffer.
 while (numberOfRequestedOverdraftMemorySegments > 0
 && numberOfRequestedMemorySegments < currentPoolSize) {
 numberOfRequestedOverdraftMemorySegments--;
 numberOfRequestedMemorySegments++;
 }
 
+// If pool size decreases, try to convert ordinary buffer to 
overdraft buffer.
+while (numberOfRequestedMemorySegments > currentPoolSize) {
+numberOfRequestedMemorySegments--;
+numberOfRequestedOverdraftMemorySegments++;
+}
+
 returnExcessMemorySegments();
 
 if (isDestroyed) {
diff --git 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
index 956d55e5948..6c0fcf13b47 100644
--- 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
+++ 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java
@@ -38,10 +38,8 @@ import java.util.ArrayDeque;
 import java.util.ArrayList;
 import java.util.Deque;
 import java.util.HashMap;
-import java.util.LinkedList;
 import java.util.List;
 import java.util.Map;
-import java.util.Queue;
 import java.util.concurrent.Callable;
 import java.util.concurrent.CompletableFuture;
 import java.util.concurrent.ExecutionException;
@@ -255,9 +253,38 @@ class LocalBufferPoolTest {
 void testDecreasePoolSize() throws Exception {
 final int maxMemorySegments = 10;
 final int requiredMemorySegments = 4;
-final int maxOverdraftBuffers = 2;
-final int largePoolSize = 5;
-final int smallPoolSize = 4;
+
+// requested buffers is equal to small pool size.
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 7, 5, 2, 5, 0, 5, 
0);
+// requested buffers is less than small pool size.
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 6, 4, 2, 2, 0, 3, 
1);
+// exceed buffers is equal to maxOverdraftBuffers
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 7, 5, 2, 7, 2, 5, 
0);
+// exceed buffers is greater than maxOverdraftBuffers
+testDecreasePoolSizeInternal(
+maxMemorySegments, requiredMemorySegments, 9, 5, 3, 9, 4, 5, 
0);
+// exceed buffers is less than 

[flink-web] 02/02: Rebuild website

2023-04-17 Thread martijnvisser
This is an automated email from the ASF dual-hosted git repository.

martijnvisser pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git

commit a77def222a81251be8d673bf12ecc3954f642c91
Author: Martijn Visser 
AuthorDate: Mon Apr 17 09:32:51 2023 +0200

Rebuild website
---
 content/community/index.html   |  2 +-
 content/getting-help/index.html|  2 +-
 content/zh/community/index.html|  2 +-
 content/zh/downloads/index.html| 12 ++--
 content/zh/getting-help/index.html |  2 +-
 5 files changed, 10 insertions(+), 10 deletions(-)

diff --git a/content/community/index.html b/content/community/index.html
index f1ac6bfbb..8008dc2e0 100644
--- a/content/community/index.html
+++ b/content/community/index.html
@@ -1591,7 +1591,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a
   Slack
   #
 
-You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache
 Flink community on Slack.
+You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>Apache
 Flink community on Slack.
 After creating an account in Slack, dont forget to introduce yourself 
in #introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the Dev mailing list.
 Any existing Slack member can also invite anyone else to join.
diff --git a/content/getting-help/index.html b/content/getting-help/index.html
index 86ab8e510..e59c8931f 100644
--- a/content/getting-help/index.html
+++ b/content/getting-help/index.html
@@ -971,7 +971,7 @@ under the License.
   Slack
   #
 
-You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache
 Flink community on Slack.
+You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>Apache
 Flink community on Slack.
 After creating an account in Slack, dont forget to introduce yourself 
in #introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the Dev 
mailing list.
 Any existing Slack member can also invite anyone else to join.
diff --git a/content/zh/community/index.html b/content/zh/community/index.html
index 99c52d2a1..dbe74d668 100644
--- a/content/zh/community/index.html
+++ b/content/zh/community/index.html
@@ -1601,7 +1601,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a
   Slack
   #
 
-你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接
+你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>此链接
 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 
规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
 在 Slack 空间交流时,请遵守以下规则:
diff --git a/content/zh/downloads/index.html b/content/zh/downloads/index.html
index ff5bc96b3..b6a7c07b9 100644
--- a/content/zh/downloads/index.html
+++ b/content/zh/downloads/index.html
@@ -910,7 +910,7 @@ https://github.com/alex-shpak/hugo-book
 Apache Flink 
Elasticsearch Connector 3.0.0
 Apache Flink JDBC 
Connector 3.0.0
 Apache Flink MongoDB 
Connector 1.0.0
-Apache Flink 
Opensearch Connector 3.0.0
+Apache Flink 
Opensearch Connector 1.0.0
 Apache Flink Pulsar 
Connector 3.0.0
 Apache Flink 
RabbitMQ Connector 3.0.0
   
@@ -1098,11 +1098,11 @@ under the License.
 
 1.16.x
 
-
-  Apache Flink Opensearch Connector 3.0.0
-  #
+
+  Apache Flink Opensearch Connector 1.0.0
+  #
 
-https://www.apache.org/dyn/closer.lua/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz;>Apache
 Flink Opensearch Connector 3.0.0 Source Release https://downloads.apache.org/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz.asc;>(asc,
 https://downloads.apache.org/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz.sha512;>sha512)
+https://www.apache.org/dyn/closer.lua/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz;>Apache
 Flink Opensearch Connector 1.0.0 Source Release https://downloads.apache.org/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz.asc;>(asc,
 https://downloads.apache.org/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz.sha512;>sha512)
 This component is compatible with Apache Flink version(s):
 
 1.16.x
@@ -1610,7 +1610,7 @@ under the License.
 Apache Flink 
Elasticsearch Connector 3.0.0
 Apache Flink JDBC 
Connector 3.0.0
 Apache Flink MongoDB 
Connector 1.0.0
-Apache Flink 
Opensearch Connector 3.0.0
+Apache Flink 
Opensearch Connector 1.0.0
 Apache Flink Pulsar 
Connector 3.0.0
 Apache Flink 
RabbitMQ Connector 

[flink-web] branch asf-site updated (b71eb1bc8 -> a77def222)

2023-04-17 Thread martijnvisser
This is an automated email from the ASF dual-hosted git repository.

martijnvisser pushed a change to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git


from b71eb1bc8 Rebuild website
 new a896bff22 [hotfix] Refresh Slack invite link
 new a77def222 Rebuild website

The 2 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails.  The revisions
listed as "add" were already present in the repository and have only
been added to this reference.


Summary of changes:
 content/community/index.html   |  2 +-
 content/getting-help/index.html|  2 +-
 content/zh/community/index.html|  2 +-
 content/zh/downloads/index.html| 12 ++--
 content/zh/getting-help/index.html |  2 +-
 docs/content.zh/community.md   |  2 +-
 docs/content.zh/getting-help.md|  2 +-
 docs/content/community.md  |  2 +-
 docs/content/getting-help.md   |  2 +-
 9 files changed, 14 insertions(+), 14 deletions(-)



[flink-web] 01/02: [hotfix] Refresh Slack invite link

2023-04-17 Thread martijnvisser
This is an automated email from the ASF dual-hosted git repository.

martijnvisser pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git

commit a896bff22862041773824058f29c10001289ba0f
Author: Martijn Visser 
AuthorDate: Mon Apr 17 09:31:29 2023 +0200

[hotfix] Refresh Slack invite link
---
 docs/content.zh/community.md| 2 +-
 docs/content.zh/getting-help.md | 2 +-
 docs/content/community.md   | 2 +-
 docs/content/getting-help.md| 2 +-
 4 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/content.zh/community.md b/docs/content.zh/community.md
index c20ce2cdb..499bde629 100644
--- a/docs/content.zh/community.md
+++ b/docs/content.zh/community.md
@@ -145,7 +145,7 @@ under the License.
 
 ## Slack
 
-你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
+你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw)
 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 
规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表](#mailing-lists)。 
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
 
diff --git a/docs/content.zh/getting-help.md b/docs/content.zh/getting-help.md
index 44c4ca575..a6e25767b 100644
--- a/docs/content.zh/getting-help.md
+++ b/docs/content.zh/getting-help.md
@@ -47,7 +47,7 @@ Apache Flink 社区每天都会回答许多用户的问题。你可以从历史
 
 ### Slack
 
-你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
 加入 Apache Flink 社区专属的 Slack 工作空间。
+你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw)
 加入 Apache Flink 社区专属的 Slack 工作空间。
 在成功加入后,不要忘记在 #introductions 频道介绍你自己。
 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表]({{< relref "community" 
>}}#mailing-lists)。 
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
diff --git a/docs/content/community.md b/docs/content/community.md
index caae2c740..96612aebe 100644
--- a/docs/content/community.md
+++ b/docs/content/community.md
@@ -146,7 +146,7 @@ If you send us an email with a code snippet, make sure that:
 
 ## Slack
 
-You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
+You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw)
 After creating an account in Slack, don't forget to introduce yourself in 
#introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the [Dev mailing list](#mailing-lists).
 Any existing Slack member can also invite anyone else to join.
diff --git a/docs/content/getting-help.md b/docs/content/getting-help.md
index 2ab1df324..4059012a3 100644
--- a/docs/content/getting-help.md
+++ b/docs/content/getting-help.md
@@ -47,7 +47,7 @@ Please note that you won't receive a response to your mail if 
you are not subscr
 
 ### Slack
 
-You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
+You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw)
 After creating an account in Slack, don't forget to introduce yourself in 
#introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the [Dev mailing list]({{< relref "community" 
>}}#mailing-lists).
 Any existing Slack member can also invite anyone else to join.



[flink-web] branch asf-site updated: Rebuild website

2023-04-17 Thread tangyun
This is an automated email from the ASF dual-hosted git repository.

tangyun pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new b71eb1bc8 Rebuild website
b71eb1bc8 is described below

commit b71eb1bc8baa251a7d1761046068b3368ec01d29
Author: Yun Tang 
AuthorDate: Mon Apr 17 14:09:15 2023 +0800

Rebuild website
---
 content/community/index.html   | 2 +-
 content/getting-help/index.html| 2 +-
 content/zh/community/index.html| 2 +-
 content/zh/getting-help/index.html | 2 +-
 4 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/content/community/index.html b/content/community/index.html
index 0db8b0c11..f1ac6bfbb 100644
--- a/content/community/index.html
+++ b/content/community/index.html
@@ -1591,7 +1591,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a
   Slack
   #
 
-You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>Apache
 Flink community on Slack.
+You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache
 Flink community on Slack.
 After creating an account in Slack, dont forget to introduce yourself 
in #introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the Dev mailing list.
 Any existing Slack member can also invite anyone else to join.
diff --git a/content/getting-help/index.html b/content/getting-help/index.html
index 22d1132ec..86ab8e510 100644
--- a/content/getting-help/index.html
+++ b/content/getting-help/index.html
@@ -971,7 +971,7 @@ under the License.
   Slack
   #
 
-You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>Apache
 Flink community on Slack.
+You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache
 Flink community on Slack.
 After creating an account in Slack, dont forget to introduce yourself 
in #introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the Dev 
mailing list.
 Any existing Slack member can also invite anyone else to join.
diff --git a/content/zh/community/index.html b/content/zh/community/index.html
index 9f1b7a6b9..99c52d2a1 100644
--- a/content/zh/community/index.html
+++ b/content/zh/community/index.html
@@ -1601,7 +1601,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a
   Slack
   #
 
-你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>此链接
+你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接
 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 
规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
 在 Slack 空间交流时,请遵守以下规则:
diff --git a/content/zh/getting-help/index.html 
b/content/zh/getting-help/index.html
index abdb62663..e82973973 100644
--- a/content/zh/getting-help/index.html
+++ b/content/zh/getting-help/index.html
@@ -981,7 +981,7 @@ under the License.
   Slack
   #
 
-你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>此链接
 加入 Apache Flink 社区专属的 Slack 工作空间。
+你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接
 加入 Apache Flink 社区专属的 Slack 工作空间。
 在成功加入后,不要忘记在 #introductions 频道介绍你自己。
 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。



[flink-web] branch asf-site updated: Update Slack invitation link

2023-04-17 Thread tangyun
This is an automated email from the ASF dual-hosted git repository.

tangyun pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/flink-web.git


The following commit(s) were added to refs/heads/asf-site by this push:
 new 2e47b9954 Update Slack invitation link
2e47b9954 is described below

commit 2e47b9954f274b765525d9e7a93e8399140d8884
Author: Yun Tang 
AuthorDate: Mon Apr 17 13:01:34 2023 +0800

Update Slack invitation link
---
 docs/content.zh/community.md| 2 +-
 docs/content.zh/getting-help.md | 2 +-
 docs/content/community.md   | 2 +-
 docs/content/getting-help.md| 2 +-
 4 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/content.zh/community.md b/docs/content.zh/community.md
index a2f78300d..c20ce2cdb 100644
--- a/docs/content.zh/community.md
+++ b/docs/content.zh/community.md
@@ -145,7 +145,7 @@ under the License.
 
 ## Slack
 
-你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q)
+你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 
规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表](#mailing-lists)。 
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
 
diff --git a/docs/content.zh/getting-help.md b/docs/content.zh/getting-help.md
index 92477b614..44c4ca575 100644
--- a/docs/content.zh/getting-help.md
+++ b/docs/content.zh/getting-help.md
@@ -47,7 +47,7 @@ Apache Flink 社区每天都会回答许多用户的问题。你可以从历史
 
 ### Slack
 
-你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q)
 加入 Apache Flink 社区专属的 Slack 工作空间。
+你可以通过 
[此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
 加入 Apache Flink 社区专属的 Slack 工作空间。
 在成功加入后,不要忘记在 #introductions 频道介绍你自己。
 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表]({{< relref "community" 
>}}#mailing-lists)。 
 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
diff --git a/docs/content/community.md b/docs/content/community.md
index 506c2ac61..caae2c740 100644
--- a/docs/content/community.md
+++ b/docs/content/community.md
@@ -146,7 +146,7 @@ If you send us an email with a code snippet, make sure that:
 
 ## Slack
 
-You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q)
+You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
 After creating an account in Slack, don't forget to introduce yourself in 
#introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the [Dev mailing list](#mailing-lists).
 Any existing Slack member can also invite anyone else to join.
diff --git a/docs/content/getting-help.md b/docs/content/getting-help.md
index 1d5fa559a..2ab1df324 100644
--- a/docs/content/getting-help.md
+++ b/docs/content/getting-help.md
@@ -47,7 +47,7 @@ Please note that you won't receive a response to your mail if 
you are not subscr
 
 ### Slack
 
-You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q)
+You can join the [Apache Flink community on 
Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA)
 After creating an account in Slack, don't forget to introduce yourself in 
#introductions.
 Due to Slack limitations the invite link expires after 100 invites. If it is 
expired, please reach out to the [Dev mailing list]({{< relref "community" 
>}}#mailing-lists).
 Any existing Slack member can also invite anyone else to join.