[flink-ml] branch master updated: [FLINK-31255] OperatorUtils#createWrappedOperatorConfig should update input and sideOutput serializers
This is an automated email from the ASF dual-hosted git repository. lindong pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/flink-ml.git The following commit(s) were added to refs/heads/master by this push: new 89468545 [FLINK-31255] OperatorUtils#createWrappedOperatorConfig should update input and sideOutput serializers 89468545 is described below commit 894685455d1c26fd45198857b7a96ee850725a59 Author: JiangXin AuthorDate: Tue Apr 18 13:47:52 2023 +0800 [FLINK-31255] OperatorUtils#createWrappedOperatorConfig should update input and sideOutput serializers This closes #229. --- .../flink/iteration/operator/OperatorUtils.java| 65 +++-- flink-ml-tests/pom.xml | 7 +++ .../iteration/UnboundedStreamIterationITCase.java | 67 ++ 3 files changed, 135 insertions(+), 4 deletions(-) diff --git a/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java b/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java index b3b629fe..3d67b7fa 100644 --- a/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java +++ b/flink-ml-iteration/src/main/java/org/apache/flink/iteration/operator/OperatorUtils.java @@ -18,21 +18,26 @@ package org.apache.flink.iteration.operator; +import org.apache.flink.api.common.typeutils.TypeSerializer; import org.apache.flink.api.java.functions.KeySelector; import org.apache.flink.configuration.Configuration; import org.apache.flink.core.fs.Path; import org.apache.flink.iteration.IterationID; import org.apache.flink.iteration.config.IterationOptions; import org.apache.flink.iteration.proxy.ProxyKeySelector; +import org.apache.flink.iteration.typeinfo.IterationRecordSerializer; +import org.apache.flink.iteration.typeinfo.IterationRecordTypeInfo; import org.apache.flink.iteration.utils.ReflectionUtils; import org.apache.flink.runtime.jobgraph.OperatorID; import org.apache.flink.statefun.flink.core.feedback.FeedbackChannel; import org.apache.flink.statefun.flink.core.feedback.FeedbackConsumer; import org.apache.flink.statefun.flink.core.feedback.FeedbackKey; import org.apache.flink.streaming.api.graph.StreamConfig; +import org.apache.flink.streaming.api.graph.StreamConfig.NetworkInputConfig; import org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator; import org.apache.flink.streaming.api.operators.StreamOperator; import org.apache.flink.util.ExceptionUtils; +import org.apache.flink.util.OutputTag; import org.apache.flink.util.function.SupplierWithException; import org.apache.flink.util.function.ThrowingConsumer; @@ -42,6 +47,7 @@ import java.util.Arrays; import java.util.Random; import java.util.UUID; import java.util.concurrent.Executor; +import java.util.stream.Stream; import static org.apache.flink.util.Preconditions.checkState; @@ -89,11 +95,10 @@ public class OperatorUtils { } } -public static StreamConfig createWrappedOperatorConfig( -StreamConfig wrapperConfig, ClassLoader cl) { -StreamConfig wrappedConfig = new StreamConfig(wrapperConfig.getConfiguration().clone()); +public static StreamConfig createWrappedOperatorConfig(StreamConfig config, ClassLoader cl) { +StreamConfig wrappedConfig = new StreamConfig(config.getConfiguration().clone()); for (int i = 0; i < wrappedConfig.getNumberOfNetworkInputs(); ++i) { -KeySelector keySelector = wrapperConfig.getStatePartitioner(i, cl); +KeySelector keySelector = config.getStatePartitioner(i, cl); if (keySelector != null) { checkState( keySelector instanceof ProxyKeySelector, @@ -104,6 +109,58 @@ public class OperatorUtils { } } +StreamConfig.InputConfig[] inputs = config.getInputs(cl); +for (int i = 0; i < inputs.length; ++i) { +if (inputs[i] instanceof NetworkInputConfig) { +TypeSerializer typeSerializerIn = +((NetworkInputConfig) inputs[i]).getTypeSerializer(); +checkState( +typeSerializerIn instanceof IterationRecordSerializer, +"The serializer of input[%s] should be IterationRecordSerializer but it is %s.", +i, +typeSerializerIn); +inputs[i] = +new NetworkInputConfig( +((IterationRecordSerializer) typeSerializerIn) +.getInnerSerializer(), +i); +} +} +wrappedConfig.setInputs(inputs); + +TypeSerializer typeSerializerOut = config.getTypeSerializerOut(cl); +checkState( +typeSerializerOut instanceof IterationRecordSerializer, +
[flink-connector-jdbc] 03/03: [FLINK-31649] Fix archunit connectors options must be public
This is an automated email from the ASF dual-hosted git repository. wanglijie pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git commit 396f205bcbed133d1e7dc0d6c3ce9487dcbc1dd7 Author: Joao Boto AuthorDate: Thu Apr 13 15:00:33 2023 +0200 [FLINK-31649] Fix archunit connectors options must be public --- .../b10c194a-0517-4744-830f-81abfb71ab18 | 2 -- .../connector/jdbc/internal/JdbcOutputFormat.java | 6 ++-- ...ons.java => InternalJdbcConnectionOptions.java} | 14 .../jdbc/table/JdbcDynamicTableFactory.java| 13 .../connector/jdbc/table/JdbcDynamicTableSink.java | 6 ++-- .../jdbc/table/JdbcDynamicTableSource.java | 6 ++-- .../jdbc/table/JdbcOutputFormatBuilder.java| 6 ++-- .../jdbc/table/JdbcRowDataLookupFunction.java | 4 +-- .../connector/jdbc/internal/JdbcFullTest.java | 10 +++--- .../jdbc/internal/JdbcTableOutputFormatTest.java | 12 +++ .../jdbc/table/JdbcAppendOnlyWriterTest.java | 4 +-- .../jdbc/table/JdbcDynamicTableFactoryTest.java| 30 - .../connector/jdbc/table/JdbcOutputFormatTest.java | 38 +++--- .../jdbc/table/JdbcRowDataInputFormatTest.java | 4 +-- .../jdbc/table/JdbcRowDataLookupFunctionTest.java | 6 ++-- 15 files changed, 80 insertions(+), 81 deletions(-) diff --git a/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18 b/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18 index 400c942..e69de29 100644 --- a/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18 +++ b/flink-connector-jdbc/archunit-violations/b10c194a-0517-4744-830f-81abfb71ab18 @@ -1,2 +0,0 @@ -Class does not reside in a package 'org.apache.flink..table' in (JdbcConnectorOptions.java:0) -org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions does not satisfy: annotated with @PublicEvolving or annotated with @Public \ No newline at end of file diff --git a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java index a869f6c..1ca8b63 100644 --- a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java +++ b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/JdbcOutputFormat.java @@ -33,7 +33,7 @@ import org.apache.flink.connector.jdbc.JdbcStatementBuilder; import org.apache.flink.connector.jdbc.internal.connection.JdbcConnectionProvider; import org.apache.flink.connector.jdbc.internal.connection.SimpleJdbcConnectionProvider; import org.apache.flink.connector.jdbc.internal.executor.JdbcBatchStatementExecutor; -import org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions; +import org.apache.flink.connector.jdbc.internal.options.InternalJdbcConnectionOptions; import org.apache.flink.connector.jdbc.internal.options.JdbcDmlOptions; import org.apache.flink.connector.jdbc.statement.FieldNamedPreparedStatementImpl; import org.apache.flink.connector.jdbc.utils.JdbcUtils; @@ -284,7 +284,7 @@ public class JdbcOutputFormat new NullPointerException("No driverName supplied.")); } -return new JdbcConnectorOptions( +return new InternalJdbcConnectionOptions( dialect.appendDefaultUrlProperties(dbURL), tableName, driverName, diff --git a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java index bff5a49..3434f60 100644 --- a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java +++ b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java @@ -25,7 +25,7 @@ import org.apache.flink.configuration.ReadableConfig; import org.apache.flink.connector.jdbc.JdbcExecutionOptions; import org.apache.flink.connector.jdbc.dialect.JdbcDialect; import org.apache.flink.connector.jdbc.dialect.JdbcDialectLoader; -import org.apache.flink.connector.jdbc.internal.options.JdbcConnectorOptions; +import org.apache.flink.connector.jdbc.internal.options.InternalJdbcConnectionOptions; import org.apache.flink.connector.jdbc.internal.options.JdbcDmlOptions; import org.apache.flink.connector.jdbc.internal.options.JdbcReadOptions; import org.apache.flink.table.connector.sink.DynamicTableSink; @@ -90,7 +90,8 @@ public class JdbcDynamicTableFactory implements DynamicTableSourceFactory, Dynam validateConfigOptions(config, context.getClassLoader()); validateDataTypeWithJdbcDialect( context.getPhysicalRowDataType(), config.get(URL), context.getClassLoader()); -
[flink-connector-jdbc] branch main updated (b6c3cb2 -> 396f205)
This is an automated email from the ASF dual-hosted git repository. wanglijie pushed a change to branch main in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git from b6c3cb2 [FLINK-31759][JDBC] Using sql_connector_download_table shortcode for table document. new 508b152 [FLINK-31649] Fix archunit ITCase needs minicluster new 6397c63 [FLINK-31649] Fix archunit returning public(evolving) methods new 396f205 [FLINK-31649] Fix archunit connectors options must be public The 3 revisions listed above as "new" are entirely new to this repository and will be described in separate emails. The revisions listed as "add" were already present in the repository and have only been added to this reference. Summary of changes: .../6b9ab1b0-c14d-4667-bab5-407b81fba98b | 36 .../b10c194a-0517-4744-830f-81abfb71ab18 | 2 -- .../d45c3af5-52c6-45fd-9926-75e75e77473a | 7 .../connector/jdbc/JdbcExactlyOnceOptions.java | 1 + .../flink/connector/jdbc/JdbcExecutionOptions.java | 1 + .../flink/connector/jdbc/catalog/JdbcCatalog.java | 2 ++ .../connector/jdbc/dialect/AbstractDialect.java| 1 + .../connector/jdbc/internal/JdbcOutputFormat.java | 6 ++-- ...ons.java => InternalJdbcConnectionOptions.java} | 14 .../jdbc/table/JdbcDynamicTableFactory.java| 13 .../connector/jdbc/table/JdbcDynamicTableSink.java | 6 ++-- .../jdbc/table/JdbcDynamicTableSource.java | 6 ++-- .../jdbc/table/JdbcOutputFormatBuilder.java| 6 ++-- .../jdbc/table/JdbcRowDataLookupFunction.java | 4 +-- .../connector/jdbc/xa/JdbcXaSinkFunctionState.java | 3 ++ .../connector/jdbc/xa/XaSinkStateHandler.java | 4 +-- .../apache/flink/connector/jdbc/JdbcITCase.java| 3 +- .../jdbc/catalog/MySqlCatalogTestBase.java | 3 +- .../jdbc/catalog/PostgresCatalogTestBase.java | 3 +- .../connector/jdbc/internal/JdbcFullTest.java | 10 +++--- .../jdbc/internal/JdbcTableOutputFormatTest.java | 12 +++ ...ProviderDriverClassConcurrentLoadingITCase.java | 3 +- .../jdbc/table/JdbcAppendOnlyWriterTest.java | 4 +-- .../jdbc/table/JdbcDynamicTableFactoryTest.java| 30 - .../connector/jdbc/table/JdbcOutputFormatTest.java | 38 +++--- .../jdbc/table/JdbcRowDataInputFormatTest.java | 4 +-- .../jdbc/table/JdbcRowDataLookupFunctionTest.java | 6 ++-- .../connector/jdbc/testutils/JdbcITCaseBase.java} | 30 - 28 files changed, 111 insertions(+), 147 deletions(-) rename flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/internal/options/{JdbcConnectorOptions.java => InternalJdbcConnectionOptions.java} (94%) copy flink-connector-jdbc/src/{main/java/org/apache/flink/connector/jdbc/internal/converter/DerbyRowConverter.java => test/java/org/apache/flink/connector/jdbc/testutils/JdbcITCaseBase.java} (55%)
[flink-connector-jdbc] 02/03: [FLINK-31649] Fix archunit returning public(evolving) methods
This is an automated email from the ASF dual-hosted git repository. wanglijie pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git commit 6397c637e8d2b5c5223b769b453692b8f03ae927 Author: Joao Boto AuthorDate: Mon Apr 17 11:35:23 2023 +0200 [FLINK-31649] Fix archunit returning public(evolving) methods --- .../archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a | 7 --- .../org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java| 1 + .../java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java | 1 + .../java/org/apache/flink/connector/jdbc/catalog/JdbcCatalog.java | 2 ++ .../org/apache/flink/connector/jdbc/dialect/AbstractDialect.java | 1 + .../apache/flink/connector/jdbc/xa/JdbcXaSinkFunctionState.java| 3 +++ .../org/apache/flink/connector/jdbc/xa/XaSinkStateHandler.java | 4 ++-- 7 files changed, 10 insertions(+), 9 deletions(-) diff --git a/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a b/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a index 4f453d5..e69de29 100644 --- a/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a +++ b/flink-connector-jdbc/archunit-violations/d45c3af5-52c6-45fd-9926-75e75e77473a @@ -1,7 +0,0 @@ -org.apache.flink.connector.jdbc.JdbcExactlyOnceOptions.builder(): Returned leaf type org.apache.flink.connector.jdbc.JdbcExactlyOnceOptions$JDBCExactlyOnceOptionsBuilder does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.JdbcExecutionOptions.builder(): Returned leaf type org.apache.flink.connector.jdbc.JdbcExecutionOptions$Builder does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.catalog.JdbcCatalog.getInternal(): Returned leaf type org.apache.flink.connector.jdbc.catalog.AbstractJdbcCatalog does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.dialect.AbstractDialect.decimalPrecisionRange(): Returned leaf type org.apache.flink.connector.jdbc.dialect.AbstractDialect$Range does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.dialect.AbstractDialect.timestampPrecisionRange(): Returned leaf type org.apache.flink.connector.jdbc.dialect.AbstractDialect$Range does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.xa.XaSinkStateHandler.load(org.apache.flink.runtime.state.FunctionInitializationContext): Returned leaf type org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated -org.apache.flink.connector.jdbc.xa.XaSinkStateHandler.store(org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState): Argument leaf type org.apache.flink.connector.jdbc.xa.JdbcXaSinkFunctionState does not satisfy: reside outside of package 'org.apache.flink..' or reside in any package ['..shaded..'] or annotated with @Public or annotated with @PublicEvolving or annotated with @Deprecated \ No newline at end of file diff --git a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java index c17ed17..47332d4 100644 --- a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java +++ b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExactlyOnceOptions.java @@ -104,6 +104,7 @@ public class JdbcExactlyOnceOptions implements Serializable { } /** JDBCExactlyOnceOptionsBuilder. */ +@PublicEvolving public static class JDBCExactlyOnceOptionsBuilder { private boolean recoveredAndRollback = DEFAULT_RECOVERED_AND_ROLLBACK; private int maxCommitAttempts = DEFAULT_MAX_COMMIT_ATTEMPTS; diff --git a/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java b/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/JdbcExecutionOptions.java index 137fb7c..c19677b 100644 ---
[flink-connector-jdbc] 01/03: [FLINK-31649] Fix archunit ITCase needs minicluster
This is an automated email from the ASF dual-hosted git repository. wanglijie pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/flink-connector-jdbc.git commit 508b152bf7ee50643b6b3844f20289b6461e3c25 Author: Joao Boto AuthorDate: Tue Mar 28 17:18:35 2023 +0200 [FLINK-31649] Fix archunit ITCase needs minicluster --- .../6b9ab1b0-c14d-4667-bab5-407b81fba98b | 36 -- .../apache/flink/connector/jdbc/JdbcITCase.java| 3 +- .../jdbc/catalog/MySqlCatalogTestBase.java | 3 +- .../jdbc/catalog/PostgresCatalogTestBase.java | 3 +- ...ProviderDriverClassConcurrentLoadingITCase.java | 3 +- .../connector/jdbc/testutils/JdbcITCaseBase.java | 36 ++ 6 files changed, 44 insertions(+), 40 deletions(-) diff --git a/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b b/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b index e54d3ff..e69de29 100644 --- a/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b +++ b/flink-connector-jdbc/archunit-violations/6b9ab1b0-c14d-4667-bab5-407b81fba98b @@ -1,36 +0,0 @@ -org.apache.flink.connector.jdbc.JdbcITCase does not satisfy: only one of the following predicates match:\ -* reside in a package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type InternalMiniClusterExtension and annotated with @RegisterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type MiniClusterExtension and annotated with @RegisterExtension\ -* reside in a package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class InternalMiniClusterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class MiniClusterExtension\ - or contain any fields that are public, static, and of type MiniClusterWithClientResource and final and annotated with @ClassRule or contain any fields that is of type MiniClusterWithClientResource and public and final and not static and annotated with @Rule -org.apache.flink.connector.jdbc.catalog.MySqlCatalogITCase does not satisfy: only one of the following predicates match:\ -* reside in a package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type InternalMiniClusterExtension and annotated with @RegisterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type MiniClusterExtension and annotated with @RegisterExtension\ -* reside in a package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class InternalMiniClusterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class MiniClusterExtension\ - or contain any fields that are public, static, and of type MiniClusterWithClientResource and final and annotated with @ClassRule or contain any fields that is of type MiniClusterWithClientResource and public and final and not static and annotated with @Rule -org.apache.flink.connector.jdbc.catalog.MySql57CatalogITCase does not satisfy: only one of the following predicates match:\ -* reside in a package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type InternalMiniClusterExtension and annotated with @RegisterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type MiniClusterExtension and annotated with @RegisterExtension\ -* reside in a package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class InternalMiniClusterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class MiniClusterExtension\ - or contain any fields that are public, static, and of type MiniClusterWithClientResource and final and annotated with @ClassRule or contain any fields that is of type MiniClusterWithClientResource and public and final and not static and annotated with @Rule -org.apache.flink.connector.jdbc.catalog.MySql56CatalogITCase does not satisfy: only one of the following predicates match:\ -* reside in a package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type InternalMiniClusterExtension and annotated with @RegisterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and contain any fields that are static, final, and of type MiniClusterExtension and annotated with @RegisterExtension\ -* reside in a package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class InternalMiniClusterExtension\ -* reside outside of package 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class MiniClusterExtension\ - or contain any fields that are public, static, and of type
[flink] branch release-1.17 updated: [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions
This is an automated email from the ASF dual-hosted git repository. guoweijie pushed a commit to branch release-1.17 in repository https://gitbox.apache.org/repos/asf/flink.git The following commit(s) were added to refs/heads/release-1.17 by this push: new fa2263d1de1 [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions fa2263d1de1 is described below commit fa2263d1de1789a5200f909333cfe2dc30104b2f Author: L AuthorDate: Tue Apr 18 00:02:53 2023 +0900 [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions --- docs/layouts/shortcodes/generated/security_auth_kerberos_section.html | 2 +- docs/layouts/shortcodes/generated/security_configuration.html | 2 +- .../src/main/java/org/apache/flink/configuration/SecurityOptions.java | 4 ++-- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html index ad063064b49..1e4e249148f 100644 --- a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html +++ b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html @@ -12,7 +12,7 @@ security.kerberos.access.hadoopFileSystems (none) ListString -A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. +A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. security.kerberos.login.contexts diff --git a/docs/layouts/shortcodes/generated/security_configuration.html b/docs/layouts/shortcodes/generated/security_configuration.html index 5bfcc1710b3..069825bdd65 100644 --- a/docs/layouts/shortcodes/generated/security_configuration.html +++ b/docs/layouts/shortcodes/generated/security_configuration.html @@ -42,7 +42,7 @@ security.kerberos.access.hadoopFileSystems (none) ListString -A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. +A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. security.kerberos.krb5-conf.path diff --git a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java index 946894ad3f4..c9f0982e48f 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java @@ -167,8 +167,8 @@ public class SecurityOptions { .noDefaultValue() .withDeprecatedKeys("yarn.security.kerberos.additionalFileSystems") .withDescription( -"A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, " -+ "security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. " +"A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, " ++ "security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. " + "The JobManager needs to have access to these filesystems to retrieve the security tokens."); //
[flink] branch master updated: [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions
This is an automated email from the ASF dual-hosted git repository. guoweijie pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/flink.git The following commit(s) were added to refs/heads/master by this push: new fe8ff46123e [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions fe8ff46123e is described below commit fe8ff46123ee2f9278066c801228416e38e2cbd4 Author: L AuthorDate: Tue Apr 18 00:02:53 2023 +0900 [FLINK-31818][docs] Fix incorrect description of security.kerberos.access.hadoopFileSystems in SecurityOptions --- docs/layouts/shortcodes/generated/security_auth_kerberos_section.html | 2 +- docs/layouts/shortcodes/generated/security_configuration.html | 2 +- .../src/main/java/org/apache/flink/configuration/SecurityOptions.java | 4 ++-- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html index ad063064b49..1e4e249148f 100644 --- a/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html +++ b/docs/layouts/shortcodes/generated/security_auth_kerberos_section.html @@ -12,7 +12,7 @@ security.kerberos.access.hadoopFileSystems (none) ListString -A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. +A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. security.kerberos.login.contexts diff --git a/docs/layouts/shortcodes/generated/security_configuration.html b/docs/layouts/shortcodes/generated/security_configuration.html index 5bfcc1710b3..069825bdd65 100644 --- a/docs/layouts/shortcodes/generated/security_configuration.html +++ b/docs/layouts/shortcodes/generated/security_configuration.html @@ -42,7 +42,7 @@ security.kerberos.access.hadoopFileSystems (none) ListString -A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. +A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. The JobManager needs to have access to these filesystems to retrieve the security tokens. security.kerberos.krb5-conf.path diff --git a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java index 946894ad3f4..c9f0982e48f 100644 --- a/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java +++ b/flink-core/src/main/java/org/apache/flink/configuration/SecurityOptions.java @@ -167,8 +167,8 @@ public class SecurityOptions { .noDefaultValue() .withDeprecatedKeys("yarn.security.kerberos.additionalFileSystems") .withDescription( -"A comma-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, " -+ "security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002,hdfs://namenode3:9003. " +"A semicolon-separated list of Kerberos-secured Hadoop filesystems Flink is going to access. For example, " ++ "security.kerberos.access.hadoopFileSystems=hdfs://namenode2:9002;hdfs://namenode3:9003. " + "The JobManager needs to have access to these filesystems to retrieve the security tokens."); //
[flink] branch master updated: [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased
This is an automated email from the ASF dual-hosted git repository. guoweijie pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/flink.git The following commit(s) were added to refs/heads/master by this push: new ffc6f3bfabd [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased ffc6f3bfabd is described below commit ffc6f3bfabd22b49b08f027400c194a8e7c9c51a Author: Weijie Guo AuthorDate: Tue Apr 11 23:33:57 2023 +0800 [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased --- .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++- .../io/network/buffer/LocalBufferPoolTest.java | 85 -- 2 files changed, 59 insertions(+), 40 deletions(-) diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java index 87b8a11e2aa..6506ab9f942 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java @@ -49,7 +49,11 @@ import static org.apache.flink.util.concurrent.FutureUtils.assertNoException; * * The size of this pool can be dynamically changed at runtime ({@link #setNumBuffers(int)}. It * will then lazily return the required number of buffers to the {@link NetworkBufferPool} to match - * its new size. + * its new size. New buffers can be requested only when {@code numberOfRequestedMemorySegments + + * numberOfRequestedOverdraftMemorySegments < currentPoolSize + maxOverdraftBuffersPerGate}. In + * order to meet this requirement, when the size of this pool changes, + * numberOfRequestedMemorySegments and numberOfRequestedOverdraftMemorySegments can be converted to + * each other. * * Availability is defined as returning a non-overdraft segment on a subsequent {@link * #requestBuffer()}/ {@link #requestBufferBuilder()} and heaving a non-blocking {@link @@ -671,13 +675,19 @@ class LocalBufferPool implements BufferPool { currentPoolSize = Math.min(numBuffers, maxNumberOfMemorySegments); -// reset overdraft buffers +// If pool size increases, try to convert overdraft buffer to ordinary buffer. while (numberOfRequestedOverdraftMemorySegments > 0 && numberOfRequestedMemorySegments < currentPoolSize) { numberOfRequestedOverdraftMemorySegments--; numberOfRequestedMemorySegments++; } +// If pool size decreases, try to convert ordinary buffer to overdraft buffer. +while (numberOfRequestedMemorySegments > currentPoolSize) { +numberOfRequestedMemorySegments--; +numberOfRequestedOverdraftMemorySegments++; +} + returnExcessMemorySegments(); if (isDestroyed) { diff --git a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java index 956d55e5948..6c0fcf13b47 100644 --- a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java +++ b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java @@ -38,10 +38,8 @@ import java.util.ArrayDeque; import java.util.ArrayList; import java.util.Deque; import java.util.HashMap; -import java.util.LinkedList; import java.util.List; import java.util.Map; -import java.util.Queue; import java.util.concurrent.Callable; import java.util.concurrent.CompletableFuture; import java.util.concurrent.ExecutionException; @@ -255,9 +253,38 @@ class LocalBufferPoolTest { void testDecreasePoolSize() throws Exception { final int maxMemorySegments = 10; final int requiredMemorySegments = 4; -final int maxOverdraftBuffers = 2; -final int largePoolSize = 5; -final int smallPoolSize = 4; + +// requested buffers is equal to small pool size. +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 7, 5, 2, 5, 0, 5, 0); +// requested buffers is less than small pool size. +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 6, 4, 2, 2, 0, 3, 1); +// exceed buffers is equal to maxOverdraftBuffers +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 7, 5, 2, 7, 2, 5, 0); +// exceed buffers is greater than maxOverdraftBuffers +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 9, 5, 3, 9, 4, 5, 0); +// exceed buffers is less than maxOverdraftBuffers +
[flink] branch release-1.16 updated (09e8e57a100 -> ab2ba9612a6)
This is an automated email from the ASF dual-hosted git repository. guoweijie pushed a change to branch release-1.16 in repository https://gitbox.apache.org/repos/asf/flink.git from 09e8e57a100 [FLINK-31779][docs] Track stable branch of externalized connector instead of specific release tag add ab2ba9612a6 [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased No new revisions were added by this update. Summary of changes: .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++- .../io/network/buffer/LocalBufferPoolTest.java | 85 -- 2 files changed, 59 insertions(+), 40 deletions(-)
[flink] branch release-1.17 updated: [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased
This is an automated email from the ASF dual-hosted git repository. guoweijie pushed a commit to branch release-1.17 in repository https://gitbox.apache.org/repos/asf/flink.git The following commit(s) were added to refs/heads/release-1.17 by this push: new be0f9293c2c [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased be0f9293c2c is described below commit be0f9293c2ce00465154ef03f7cef29dd3116b8e Author: Weijie Guo AuthorDate: Tue Apr 11 23:33:57 2023 +0800 [FLINK-31763][runtime] Convert requested buffers to overdraft buffers when pool size is decreased --- .../runtime/io/network/buffer/LocalBufferPool.java | 14 +++- .../io/network/buffer/LocalBufferPoolTest.java | 85 -- 2 files changed, 59 insertions(+), 40 deletions(-) diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java index 87b8a11e2aa..6506ab9f942 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPool.java @@ -49,7 +49,11 @@ import static org.apache.flink.util.concurrent.FutureUtils.assertNoException; * * The size of this pool can be dynamically changed at runtime ({@link #setNumBuffers(int)}. It * will then lazily return the required number of buffers to the {@link NetworkBufferPool} to match - * its new size. + * its new size. New buffers can be requested only when {@code numberOfRequestedMemorySegments + + * numberOfRequestedOverdraftMemorySegments < currentPoolSize + maxOverdraftBuffersPerGate}. In + * order to meet this requirement, when the size of this pool changes, + * numberOfRequestedMemorySegments and numberOfRequestedOverdraftMemorySegments can be converted to + * each other. * * Availability is defined as returning a non-overdraft segment on a subsequent {@link * #requestBuffer()}/ {@link #requestBufferBuilder()} and heaving a non-blocking {@link @@ -671,13 +675,19 @@ class LocalBufferPool implements BufferPool { currentPoolSize = Math.min(numBuffers, maxNumberOfMemorySegments); -// reset overdraft buffers +// If pool size increases, try to convert overdraft buffer to ordinary buffer. while (numberOfRequestedOverdraftMemorySegments > 0 && numberOfRequestedMemorySegments < currentPoolSize) { numberOfRequestedOverdraftMemorySegments--; numberOfRequestedMemorySegments++; } +// If pool size decreases, try to convert ordinary buffer to overdraft buffer. +while (numberOfRequestedMemorySegments > currentPoolSize) { +numberOfRequestedMemorySegments--; +numberOfRequestedOverdraftMemorySegments++; +} + returnExcessMemorySegments(); if (isDestroyed) { diff --git a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java index 956d55e5948..6c0fcf13b47 100644 --- a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java +++ b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/buffer/LocalBufferPoolTest.java @@ -38,10 +38,8 @@ import java.util.ArrayDeque; import java.util.ArrayList; import java.util.Deque; import java.util.HashMap; -import java.util.LinkedList; import java.util.List; import java.util.Map; -import java.util.Queue; import java.util.concurrent.Callable; import java.util.concurrent.CompletableFuture; import java.util.concurrent.ExecutionException; @@ -255,9 +253,38 @@ class LocalBufferPoolTest { void testDecreasePoolSize() throws Exception { final int maxMemorySegments = 10; final int requiredMemorySegments = 4; -final int maxOverdraftBuffers = 2; -final int largePoolSize = 5; -final int smallPoolSize = 4; + +// requested buffers is equal to small pool size. +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 7, 5, 2, 5, 0, 5, 0); +// requested buffers is less than small pool size. +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 6, 4, 2, 2, 0, 3, 1); +// exceed buffers is equal to maxOverdraftBuffers +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 7, 5, 2, 7, 2, 5, 0); +// exceed buffers is greater than maxOverdraftBuffers +testDecreasePoolSizeInternal( +maxMemorySegments, requiredMemorySegments, 9, 5, 3, 9, 4, 5, 0); +// exceed buffers is less than
[flink-web] 02/02: Rebuild website
This is an automated email from the ASF dual-hosted git repository. martijnvisser pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/flink-web.git commit a77def222a81251be8d673bf12ecc3954f642c91 Author: Martijn Visser AuthorDate: Mon Apr 17 09:32:51 2023 +0200 Rebuild website --- content/community/index.html | 2 +- content/getting-help/index.html| 2 +- content/zh/community/index.html| 2 +- content/zh/downloads/index.html| 12 ++-- content/zh/getting-help/index.html | 2 +- 5 files changed, 10 insertions(+), 10 deletions(-) diff --git a/content/community/index.html b/content/community/index.html index f1ac6bfbb..8008dc2e0 100644 --- a/content/community/index.html +++ b/content/community/index.html @@ -1591,7 +1591,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a Slack # -You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache Flink community on Slack. +You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>Apache Flink community on Slack. After creating an account in Slack, dont forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the Dev mailing list. Any existing Slack member can also invite anyone else to join. diff --git a/content/getting-help/index.html b/content/getting-help/index.html index 86ab8e510..e59c8931f 100644 --- a/content/getting-help/index.html +++ b/content/getting-help/index.html @@ -971,7 +971,7 @@ under the License. Slack # -You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache Flink community on Slack. +You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>Apache Flink community on Slack. After creating an account in Slack, dont forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the Dev mailing list. Any existing Slack member can also invite anyone else to join. diff --git a/content/zh/community/index.html b/content/zh/community/index.html index 99c52d2a1..dbe74d668 100644 --- a/content/zh/community/index.html +++ b/content/zh/community/index.html @@ -1601,7 +1601,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a Slack # -你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接 +你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw;>此链接 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 在 Slack 空间交流时,请遵守以下规则: diff --git a/content/zh/downloads/index.html b/content/zh/downloads/index.html index ff5bc96b3..b6a7c07b9 100644 --- a/content/zh/downloads/index.html +++ b/content/zh/downloads/index.html @@ -910,7 +910,7 @@ https://github.com/alex-shpak/hugo-book Apache Flink Elasticsearch Connector 3.0.0 Apache Flink JDBC Connector 3.0.0 Apache Flink MongoDB Connector 1.0.0 -Apache Flink Opensearch Connector 3.0.0 +Apache Flink Opensearch Connector 1.0.0 Apache Flink Pulsar Connector 3.0.0 Apache Flink RabbitMQ Connector 3.0.0 @@ -1098,11 +1098,11 @@ under the License. 1.16.x - - Apache Flink Opensearch Connector 3.0.0 - # + + Apache Flink Opensearch Connector 1.0.0 + # -https://www.apache.org/dyn/closer.lua/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz;>Apache Flink Opensearch Connector 3.0.0 Source Release https://downloads.apache.org/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz.asc;>(asc, https://downloads.apache.org/flink/flink-connector-opensearch-3.0.0/flink-connector-opensearch-3.0.0-src.tgz.sha512;>sha512) +https://www.apache.org/dyn/closer.lua/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz;>Apache Flink Opensearch Connector 1.0.0 Source Release https://downloads.apache.org/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz.asc;>(asc, https://downloads.apache.org/flink/flink-connector-opensearch-1.0.0/flink-connector-opensearch-1.0.0-src.tgz.sha512;>sha512) This component is compatible with Apache Flink version(s): 1.16.x @@ -1610,7 +1610,7 @@ under the License. Apache Flink Elasticsearch Connector 3.0.0 Apache Flink JDBC Connector 3.0.0 Apache Flink MongoDB Connector 1.0.0 -Apache Flink Opensearch Connector 3.0.0 +Apache Flink Opensearch Connector 1.0.0 Apache Flink Pulsar Connector 3.0.0 Apache Flink RabbitMQ Connector
[flink-web] branch asf-site updated (b71eb1bc8 -> a77def222)
This is an automated email from the ASF dual-hosted git repository. martijnvisser pushed a change to branch asf-site in repository https://gitbox.apache.org/repos/asf/flink-web.git from b71eb1bc8 Rebuild website new a896bff22 [hotfix] Refresh Slack invite link new a77def222 Rebuild website The 2 revisions listed above as "new" are entirely new to this repository and will be described in separate emails. The revisions listed as "add" were already present in the repository and have only been added to this reference. Summary of changes: content/community/index.html | 2 +- content/getting-help/index.html| 2 +- content/zh/community/index.html| 2 +- content/zh/downloads/index.html| 12 ++-- content/zh/getting-help/index.html | 2 +- docs/content.zh/community.md | 2 +- docs/content.zh/getting-help.md| 2 +- docs/content/community.md | 2 +- docs/content/getting-help.md | 2 +- 9 files changed, 14 insertions(+), 14 deletions(-)
[flink-web] 01/02: [hotfix] Refresh Slack invite link
This is an automated email from the ASF dual-hosted git repository. martijnvisser pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/flink-web.git commit a896bff22862041773824058f29c10001289ba0f Author: Martijn Visser AuthorDate: Mon Apr 17 09:31:29 2023 +0200 [hotfix] Refresh Slack invite link --- docs/content.zh/community.md| 2 +- docs/content.zh/getting-help.md | 2 +- docs/content/community.md | 2 +- docs/content/getting-help.md| 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/content.zh/community.md b/docs/content.zh/community.md index c20ce2cdb..499bde629 100644 --- a/docs/content.zh/community.md +++ b/docs/content.zh/community.md @@ -145,7 +145,7 @@ under the License. ## Slack -你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) +你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw) 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表](#mailing-lists)。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 diff --git a/docs/content.zh/getting-help.md b/docs/content.zh/getting-help.md index 44c4ca575..a6e25767b 100644 --- a/docs/content.zh/getting-help.md +++ b/docs/content.zh/getting-help.md @@ -47,7 +47,7 @@ Apache Flink 社区每天都会回答许多用户的问题。你可以从历史 ### Slack -你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) 加入 Apache Flink 社区专属的 Slack 工作空间。 +你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw) 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表]({{< relref "community" >}}#mailing-lists)。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 diff --git a/docs/content/community.md b/docs/content/community.md index caae2c740..96612aebe 100644 --- a/docs/content/community.md +++ b/docs/content/community.md @@ -146,7 +146,7 @@ If you send us an email with a code snippet, make sure that: ## Slack -You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) +You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw) After creating an account in Slack, don't forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the [Dev mailing list](#mailing-lists). Any existing Slack member can also invite anyone else to join. diff --git a/docs/content/getting-help.md b/docs/content/getting-help.md index 2ab1df324..4059012a3 100644 --- a/docs/content/getting-help.md +++ b/docs/content/getting-help.md @@ -47,7 +47,7 @@ Please note that you won't receive a response to your mail if you are not subscr ### Slack -You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) +You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1thin01ch-tYuj6Zwu8qf0QsivHY0anw) After creating an account in Slack, don't forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the [Dev mailing list]({{< relref "community" >}}#mailing-lists). Any existing Slack member can also invite anyone else to join.
[flink-web] branch asf-site updated: Rebuild website
This is an automated email from the ASF dual-hosted git repository. tangyun pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/flink-web.git The following commit(s) were added to refs/heads/asf-site by this push: new b71eb1bc8 Rebuild website b71eb1bc8 is described below commit b71eb1bc8baa251a7d1761046068b3368ec01d29 Author: Yun Tang AuthorDate: Mon Apr 17 14:09:15 2023 +0800 Rebuild website --- content/community/index.html | 2 +- content/getting-help/index.html| 2 +- content/zh/community/index.html| 2 +- content/zh/getting-help/index.html | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/content/community/index.html b/content/community/index.html index 0db8b0c11..f1ac6bfbb 100644 --- a/content/community/index.html +++ b/content/community/index.html @@ -1591,7 +1591,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a Slack # -You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>Apache Flink community on Slack. +You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache Flink community on Slack. After creating an account in Slack, dont forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the Dev mailing list. Any existing Slack member can also invite anyone else to join. diff --git a/content/getting-help/index.html b/content/getting-help/index.html index 22d1132ec..86ab8e510 100644 --- a/content/getting-help/index.html +++ b/content/getting-help/index.html @@ -971,7 +971,7 @@ under the License. Slack # -You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>Apache Flink community on Slack. +You can join the https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>Apache Flink community on Slack. After creating an account in Slack, dont forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the Dev mailing list. Any existing Slack member can also invite anyone else to join. diff --git a/content/zh/community/index.html b/content/zh/community/index.html index 9f1b7a6b9..99c52d2a1 100644 --- a/content/zh/community/index.html +++ b/content/zh/community/index.html @@ -1601,7 +1601,7 @@ data-td_id='Archive'>https://lists.apache.org/list.html?commits@flink.a Slack # -你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>此链接 +你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 在 Slack 空间交流时,请遵守以下规则: diff --git a/content/zh/getting-help/index.html b/content/zh/getting-help/index.html index abdb62663..e82973973 100644 --- a/content/zh/getting-help/index.html +++ b/content/zh/getting-help/index.html @@ -981,7 +981,7 @@ under the License. Slack # -你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q;>此链接 加入 Apache Flink 社区专属的 Slack 工作空间。 +你可以通过 https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA;>此链接 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 Dev 邮件列表。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。
[flink-web] branch asf-site updated: Update Slack invitation link
This is an automated email from the ASF dual-hosted git repository. tangyun pushed a commit to branch asf-site in repository https://gitbox.apache.org/repos/asf/flink-web.git The following commit(s) were added to refs/heads/asf-site by this push: new 2e47b9954 Update Slack invitation link 2e47b9954 is described below commit 2e47b9954f274b765525d9e7a93e8399140d8884 Author: Yun Tang AuthorDate: Mon Apr 17 13:01:34 2023 +0800 Update Slack invitation link --- docs/content.zh/community.md| 2 +- docs/content.zh/getting-help.md | 2 +- docs/content/community.md | 2 +- docs/content/getting-help.md| 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/docs/content.zh/community.md b/docs/content.zh/community.md index a2f78300d..c20ce2cdb 100644 --- a/docs/content.zh/community.md +++ b/docs/content.zh/community.md @@ -145,7 +145,7 @@ under the License. ## Slack -你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q) +你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表](#mailing-lists)。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 diff --git a/docs/content.zh/getting-help.md b/docs/content.zh/getting-help.md index 92477b614..44c4ca575 100644 --- a/docs/content.zh/getting-help.md +++ b/docs/content.zh/getting-help.md @@ -47,7 +47,7 @@ Apache Flink 社区每天都会回答许多用户的问题。你可以从历史 ### Slack -你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q) 加入 Apache Flink 社区专属的 Slack 工作空间。 +你可以通过 [此链接](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) 加入 Apache Flink 社区专属的 Slack 工作空间。 在成功加入后,不要忘记在 #introductions 频道介绍你自己。 Slack 规定每个邀请链接最多可邀请 100 人,如果遇到上述链接失效的情况,请联系 [Dev 邮件列表]({{< relref "community" >}}#mailing-lists)。 所有已经加入社区 Slack 空间的成员同样可以邀请新成员加入。 diff --git a/docs/content/community.md b/docs/content/community.md index 506c2ac61..caae2c740 100644 --- a/docs/content/community.md +++ b/docs/content/community.md @@ -146,7 +146,7 @@ If you send us an email with a code snippet, make sure that: ## Slack -You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q) +You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) After creating an account in Slack, don't forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the [Dev mailing list](#mailing-lists). Any existing Slack member can also invite anyone else to join. diff --git a/docs/content/getting-help.md b/docs/content/getting-help.md index 1d5fa559a..2ab1df324 100644 --- a/docs/content/getting-help.md +++ b/docs/content/getting-help.md @@ -47,7 +47,7 @@ Please note that you won't receive a response to your mail if you are not subscr ### Slack -You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1qxczhnzb-Redy4vcPAiTkfcBw81Dm8Q) +You can join the [Apache Flink community on Slack.](https://join.slack.com/t/apache-flink/shared_invite/zt-1ta0su2np-lCCV6xD7XeKjwQuHMOTBIA) After creating an account in Slack, don't forget to introduce yourself in #introductions. Due to Slack limitations the invite link expires after 100 invites. If it is expired, please reach out to the [Dev mailing list]({{< relref "community" >}}#mailing-lists). Any existing Slack member can also invite anyone else to join.