This is an automated email from the ASF dual-hosted git repository. rmetzger pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/flink.git
The following commit(s) were added to refs/heads/master by this push: new b78cd3f5363 [FLINK-30636] [docs]: Typo fix; 'to to' -> 'to' b78cd3f5363 is described below commit b78cd3f536331d02eff9af4702904f331d90bc07 Author: Gunnar Morling <gunnar.morl...@googlemail.com> AuthorDate: Wed Jan 11 11:22:26 2023 +0100 [FLINK-30636] [docs]: Typo fix; 'to to' -> 'to' --- .../docs/dev/table/hive-compatibility/hive-dialect/insert.md | 2 +- docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md | 2 +- .../src/main/java/org/apache/flink/api/common/state/ListState.java | 2 +- flink-python/pyflink/datastream/state.py | 2 +- .../org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java | 2 +- .../src/main/java/org/apache/flink/runtime/execution/Environment.java | 2 +- .../org/apache/flink/runtime/io/network/api/TaskEventHandler.java | 3 +-- .../org/apache/flink/runtime/iterative/task/IterationHeadTask.java | 2 +- .../apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java | 4 ++-- .../org/apache/flink/runtime/state/internal/InternalListState.java | 2 +- .../runtime/io/network/partition/consumer/LocalInputChannelTest.java | 4 ++-- .../streaming/state/snapshot/RocksDBFullSnapshotResources.java | 2 +- .../streaming/api/functions/source/ContinuousFileReaderOperator.java | 2 +- .../table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java | 2 +- .../flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java | 3 +-- .../table/planner/plan/nodes/exec/common/CommonExecLegacySink.java | 2 +- .../table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java | 2 +- .../flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java | 2 +- .../planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala | 2 +- .../table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala | 4 +--- .../planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala | 2 +- .../table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala | 2 +- .../scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala | 2 +- .../test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java | 2 +- 24 files changed, 26 insertions(+), 30 deletions(-) diff --git a/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md b/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md index c4d862ef628..938ed6002e7 100644 --- a/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md +++ b/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md @@ -74,7 +74,7 @@ The dynamic partition columns must be specified last among the columns in the `S **Note:** In Hive, by default, users must specify at least one static partition in case of accidentally overwriting all partitions, and users can -set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to to allow all partitions to be dynamic. +set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to allow all partitions to be dynamic. But in Flink's Hive dialect, it'll always be `nonstrict` mode which means all partitions are allowed to be dynamic. {{< /hint >}} diff --git a/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md b/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md index c4d862ef628..938ed6002e7 100644 --- a/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md +++ b/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md @@ -74,7 +74,7 @@ The dynamic partition columns must be specified last among the columns in the `S **Note:** In Hive, by default, users must specify at least one static partition in case of accidentally overwriting all partitions, and users can -set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to to allow all partitions to be dynamic. +set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to allow all partitions to be dynamic. But in Flink's Hive dialect, it'll always be `nonstrict` mode which means all partitions are allowed to be dynamic. {{< /hint >}} diff --git a/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java b/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java index fb6492834d7..7508054f5c0 100644 --- a/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java +++ b/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java @@ -44,7 +44,7 @@ import java.util.List; public interface ListState<T> extends MergingState<T, Iterable<T>> { /** - * Updates the operator state accessible by {@link #get()} by updating existing values to to the + * Updates the operator state accessible by {@link #get()} by updating existing values to the * given list of values. The next time {@link #get()} is called (for the same state partition) * the returned state will represent the updated list. * diff --git a/flink-python/pyflink/datastream/state.py b/flink-python/pyflink/datastream/state.py index f38b9963bd6..edda52a9571 100644 --- a/flink-python/pyflink/datastream/state.py +++ b/flink-python/pyflink/datastream/state.py @@ -193,7 +193,7 @@ class ListState(MergingState[T, Iterable[T]]): @abstractmethod def update(self, values: List[T]) -> None: """ - Updating existing values to to the given list of values. + Updating existing values to the given list of values. """ pass diff --git a/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java b/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java index f4e0094c9fb..0741e3b15de 100644 --- a/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java +++ b/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java @@ -122,7 +122,7 @@ public class HttpTestClient implements AutoCloseable { } /** - * Sends a request to to the server. + * Sends a request to the server. * * <pre> * HttpRequest request = new DefaultFullHttpRequest(HttpVersion.HTTP_1_1, HttpMethod.GET, "/overview"); diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java index 3490160a606..0f365feaf78 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java @@ -188,7 +188,7 @@ public interface Environment { TaskKvStateRegistry getTaskKvStateRegistry(); /** - * Confirms that the invokable has successfully completed all steps it needed to to for the + * Confirms that the invokable has successfully completed all steps it needed to for the * checkpoint with the give checkpoint-ID. This method does not include any state in the * checkpoint. * diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java index f78b4833dbc..5fb2cb42172 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java @@ -25,8 +25,7 @@ import org.apache.flink.shaded.guava31.com.google.common.collect.HashMultimap; import org.apache.flink.shaded.guava31.com.google.common.collect.Multimap; /** - * The event handler manages {@link EventListener} instances and allows to to publish events to - * them. + * The event handler manages {@link EventListener} instances and allows to publish events to them. */ public class TaskEventHandler { diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java index 23096305053..9402a8a3acb 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java @@ -74,7 +74,7 @@ import java.util.List; * will send a {@link TerminationEvent} to all it's connected tasks, signaling them to shutdown. * * <p>Assumption on the ordering of the outputs: - The first n output gates write to channels that - * go to the tasks of the step function. - The next m output gates to to the tasks that consume the + * go to the tasks of the step function. - The next m output gates to the tasks that consume the * final solution. - The last output gate connects to the synchronization task. * * @param <X> The type of the bulk partial solution / solution set and the final output. diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java b/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java index 9fc9c4455cb..608a6a23d7f 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java @@ -30,8 +30,8 @@ import org.apache.flink.util.SerializedValue; import java.util.concurrent.CompletableFuture; /** - * Gateway to send an {@link OperatorEvent} or {@link CoordinationRequest} from a Task to to the - * {@link OperatorCoordinator} JobManager side. + * Gateway to send an {@link OperatorEvent} or {@link CoordinationRequest} from a Task to the {@link + * OperatorCoordinator} JobManager side. * * <p>This is the first step in the chain of sending Operator Events and Requests from Operator to * Coordinator. Each layer adds further context, so that the inner layers do not need to know about diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java b/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java index 8a785226f3c..bfb75a00bb2 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java @@ -35,7 +35,7 @@ public interface InternalListState<K, N, T> extends InternalMergingState<K, N, T, List<T>, Iterable<T>>, ListState<T> { /** - * Updates the operator state accessible by {@link #get()} by updating existing values to to the + * Updates the operator state accessible by {@link #get()} by updating existing values to the * given list of values. The next time {@link #get()} is called (for the same state partition) * the returned state will represent the updated list. * diff --git a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java index fcd404d5f83..6940707cc9b 100644 --- a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java +++ b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java @@ -401,8 +401,8 @@ public class LocalInputChannelTest { * and has not much more general value. If it becomes obsolete at some point (future greatness * ;)), feel free to remove it. * - * <p>The fix in the end was to to not acquire the channels lock when releasing it and/or not - * doing any input gate callbacks while holding the channel's lock. I decided to do both. + * <p>The fix in the end was to not acquire the channels lock when releasing it and/or not doing + * any input gate callbacks while holding the channel's lock. I decided to do both. */ @Test public void testConcurrentReleaseAndRetriggerPartitionRequest() throws Exception { diff --git a/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java b/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java index db58c24f01f..1cc3b6cbd91 100644 --- a/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java +++ b/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java @@ -90,7 +90,7 @@ public class RocksDBFullSnapshotResources<K> implements FullSnapshotResources<K> this.keySerializer = keySerializer; this.streamCompressionDecorator = streamCompressionDecorator; - // we need to to this in the constructor, i.e. in the synchronous part of the snapshot + // we need to do this in the constructor, i.e. in the synchronous part of the snapshot // TODO: better yet, we can do it outside the constructor this.metaData = fillMetaData(metaDataCopy); } diff --git a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java index 1616362725d..fbaabb7de9a 100644 --- a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java +++ b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java @@ -539,7 +539,7 @@ public class ContinuousFileReaderOperator<OUT, T extends TimestampedInputSplit> checkpointedState.clear(); throw new Exception( - "Could not add timestamped file input splits to to operator " + "Could not add timestamped file input splits to operator " + "state backend of operator " + getOperatorName() + '.', diff --git a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java index c4783660280..d264212c3bd 100644 --- a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java +++ b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java @@ -35,7 +35,7 @@ import javax.annotation.Nullable; import java.lang.reflect.Modifier; /** - * Batch {@link ExecNode} to to write data into an external sink defined by a {@link TableSink}. + * Batch {@link ExecNode} to write data into an external sink defined by a {@link TableSink}. * * @param <T> The return type of the {@link TableSink}. */ diff --git a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java index 25040ee88e6..9a28b1ccf7f 100644 --- a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java +++ b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java @@ -44,8 +44,7 @@ import java.util.Collections; import java.util.List; /** - * Batch {@link ExecNode} to to write data into an external sink defined by a {@link - * DynamicTableSink}. + * Batch {@link ExecNode} to write data into an external sink defined by a {@link DynamicTableSink}. */ public class BatchExecSink extends CommonExecSink implements BatchExecNode<Object> { public BatchExecSink( diff --git a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java index 96884df250f..d149f03baa0 100644 --- a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java +++ b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java @@ -57,7 +57,7 @@ import java.util.Collections; import java.util.List; /** - * Base {@link ExecNode} to to write data into an external sink defined by a {@link TableSink}. + * Base {@link ExecNode} to write data into an external sink defined by a {@link TableSink}. * * @param <T> The return type of the {@link TableSink}. */ diff --git a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java index 04fdde6c11b..82141baac1b 100644 --- a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java +++ b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java @@ -38,7 +38,7 @@ import java.util.List; import java.util.stream.Collectors; /** - * Stream {@link ExecNode} to to write data into an external sink defined by a {@link TableSink}. + * Stream {@link ExecNode} to write data into an external sink defined by a {@link TableSink}. * * @param <T> The return type of the {@link TableSink}. */ diff --git a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java index 79b85a45312..8b6c5614ee0 100644 --- a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java +++ b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java @@ -66,7 +66,7 @@ import java.util.List; import java.util.stream.Collectors; /** - * Stream {@link ExecNode} to to write data into an external sink defined by a {@link + * Stream {@link ExecNode} to write data into an external sink defined by a {@link * DynamicTableSink}. */ @ExecNodeMetadata( diff --git a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala index c9b024bc907..9b4ebe9ad35 100644 --- a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala +++ b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala @@ -32,7 +32,7 @@ import org.apache.calcite.rel.hint.RelHint import java.util /** - * Batch physical RelNode to to write data into an external sink defined by a [[TableSink]]. + * Batch physical RelNode to write data into an external sink defined by a [[TableSink]]. * * @tparam T * The return type of the [[TableSink]]. diff --git a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala index 457b8faee50..8080ec163f3 100644 --- a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala +++ b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala @@ -33,9 +33,7 @@ import org.apache.calcite.rel.hint.RelHint import java.util -/** - * Batch physical RelNode to to write data into an external sink defined by a [[DynamicTableSink]]. - */ +/** Batch physical RelNode to write data into an external sink defined by a [[DynamicTableSink]]. */ class BatchPhysicalSink( cluster: RelOptCluster, traitSet: RelTraitSet, diff --git a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala index 1408ed41cde..7a7d95c6dff 100644 --- a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala +++ b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala @@ -32,7 +32,7 @@ import org.apache.calcite.rel.hint.RelHint import java.util /** - * Stream physical RelNode to to write data into an external sink defined by a [[TableSink]]. + * Stream physical RelNode to write data into an external sink defined by a [[TableSink]]. * * @tparam T * The return type of the [[TableSink]]. diff --git a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala index 19b8c600ae7..a058c4b6e68 100644 --- a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala +++ b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala @@ -36,7 +36,7 @@ import org.apache.calcite.rel.hint.RelHint import java.util /** - * Stream physical RelNode to to write data into an external sink defined by a [[DynamicTableSink]]. + * Stream physical RelNode to write data into an external sink defined by a [[DynamicTableSink]]. */ class StreamPhysicalSink( cluster: RelOptCluster, diff --git a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala index ffc8c0f54af..fec071307f8 100644 --- a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala +++ b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala @@ -47,7 +47,7 @@ class DefaultRelShuttle extends RelHomogeneousShuttle { } /** - * Convert all [[QueryOperationCatalogViewTable]]s (including tables in [[RexSubQuery]]) to to a + * Convert all [[QueryOperationCatalogViewTable]]s (including tables in [[RexSubQuery]]) to a * relational expression. */ class ExpandTableScanShuttle extends RelShuttleImpl { diff --git a/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java b/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java index 1e074368fe0..7e4fe0aa16b 100644 --- a/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java +++ b/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java @@ -52,7 +52,7 @@ import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; /** - * This test uses a custom non-serializable data type to to ensure that state serializability is + * This test uses a custom non-serializable data type to ensure that state serializability is * handled correctly. */ @SuppressWarnings("serial")