This is an automated email from the ASF dual-hosted git repository.

rmetzger pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink.git


The following commit(s) were added to refs/heads/master by this push:
     new b78cd3f5363 [FLINK-30636] [docs]: Typo fix; 'to to' -> 'to'
b78cd3f5363 is described below

commit b78cd3f536331d02eff9af4702904f331d90bc07
Author: Gunnar Morling <gunnar.morl...@googlemail.com>
AuthorDate: Wed Jan 11 11:22:26 2023 +0100

    [FLINK-30636] [docs]: Typo fix; 'to to' -> 'to'
---
 .../docs/dev/table/hive-compatibility/hive-dialect/insert.md          | 2 +-
 docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md | 2 +-
 .../src/main/java/org/apache/flink/api/common/state/ListState.java    | 2 +-
 flink-python/pyflink/datastream/state.py                              | 2 +-
 .../org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java | 2 +-
 .../src/main/java/org/apache/flink/runtime/execution/Environment.java | 2 +-
 .../org/apache/flink/runtime/io/network/api/TaskEventHandler.java     | 3 +--
 .../org/apache/flink/runtime/iterative/task/IterationHeadTask.java    | 2 +-
 .../apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java | 4 ++--
 .../org/apache/flink/runtime/state/internal/InternalListState.java    | 2 +-
 .../runtime/io/network/partition/consumer/LocalInputChannelTest.java  | 4 ++--
 .../streaming/state/snapshot/RocksDBFullSnapshotResources.java        | 2 +-
 .../streaming/api/functions/source/ContinuousFileReaderOperator.java  | 2 +-
 .../table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java      | 2 +-
 .../flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java      | 3 +--
 .../table/planner/plan/nodes/exec/common/CommonExecLegacySink.java    | 2 +-
 .../table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java    | 2 +-
 .../flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java    | 2 +-
 .../planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala   | 2 +-
 .../table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala   | 4 +---
 .../planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala | 2 +-
 .../table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala | 2 +-
 .../scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala | 2 +-
 .../test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java   | 2 +-
 24 files changed, 26 insertions(+), 30 deletions(-)

diff --git 
a/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md 
b/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md
index c4d862ef628..938ed6002e7 100644
--- a/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md
+++ b/docs/content.zh/docs/dev/table/hive-compatibility/hive-dialect/insert.md
@@ -74,7 +74,7 @@ The dynamic partition columns must be specified last among 
the columns in the `S
 **Note:**
 
 In Hive, by default, users must specify at least one static partition in case 
of accidentally overwriting all partitions, and users can
-set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to to 
allow all partitions to be dynamic.
+set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to 
allow all partitions to be dynamic.
 
 But in Flink's Hive dialect, it'll always be `nonstrict` mode which means all 
partitions are allowed to be dynamic.
 {{< /hint >}}
diff --git 
a/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md 
b/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md
index c4d862ef628..938ed6002e7 100644
--- a/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md
+++ b/docs/content/docs/dev/table/hive-compatibility/hive-dialect/insert.md
@@ -74,7 +74,7 @@ The dynamic partition columns must be specified last among 
the columns in the `S
 **Note:**
 
 In Hive, by default, users must specify at least one static partition in case 
of accidentally overwriting all partitions, and users can
-set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to to 
allow all partitions to be dynamic.
+set the configuration `hive.exec.dynamic.partition.mode` to `nonstrict` to 
allow all partitions to be dynamic.
 
 But in Flink's Hive dialect, it'll always be `nonstrict` mode which means all 
partitions are allowed to be dynamic.
 {{< /hint >}}
diff --git 
a/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java 
b/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java
index fb6492834d7..7508054f5c0 100644
--- a/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java
+++ b/flink-core/src/main/java/org/apache/flink/api/common/state/ListState.java
@@ -44,7 +44,7 @@ import java.util.List;
 public interface ListState<T> extends MergingState<T, Iterable<T>> {
 
     /**
-     * Updates the operator state accessible by {@link #get()} by updating 
existing values to to the
+     * Updates the operator state accessible by {@link #get()} by updating 
existing values to the
      * given list of values. The next time {@link #get()} is called (for the 
same state partition)
      * the returned state will represent the updated list.
      *
diff --git a/flink-python/pyflink/datastream/state.py 
b/flink-python/pyflink/datastream/state.py
index f38b9963bd6..edda52a9571 100644
--- a/flink-python/pyflink/datastream/state.py
+++ b/flink-python/pyflink/datastream/state.py
@@ -193,7 +193,7 @@ class ListState(MergingState[T, Iterable[T]]):
     @abstractmethod
     def update(self, values: List[T]) -> None:
         """
-        Updating existing values to to the given list of values.
+        Updating existing values to the given list of values.
         """
         pass
 
diff --git 
a/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java
 
b/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java
index f4e0094c9fb..0741e3b15de 100644
--- 
a/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java
+++ 
b/flink-runtime-web/src/test/java/org/apache/flink/runtime/webmonitor/testutils/HttpTestClient.java
@@ -122,7 +122,7 @@ public class HttpTestClient implements AutoCloseable {
     }
 
     /**
-     * Sends a request to to the server.
+     * Sends a request to the server.
      *
      * <pre>
      * HttpRequest request = new DefaultFullHttpRequest(HttpVersion.HTTP_1_1, 
HttpMethod.GET, "/overview");
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java
index 3490160a606..0f365feaf78 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/Environment.java
@@ -188,7 +188,7 @@ public interface Environment {
     TaskKvStateRegistry getTaskKvStateRegistry();
 
     /**
-     * Confirms that the invokable has successfully completed all steps it 
needed to to for the
+     * Confirms that the invokable has successfully completed all steps it 
needed to for the
      * checkpoint with the give checkpoint-ID. This method does not include 
any state in the
      * checkpoint.
      *
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java
index f78b4833dbc..5fb2cb42172 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/api/TaskEventHandler.java
@@ -25,8 +25,7 @@ import 
org.apache.flink.shaded.guava31.com.google.common.collect.HashMultimap;
 import org.apache.flink.shaded.guava31.com.google.common.collect.Multimap;
 
 /**
- * The event handler manages {@link EventListener} instances and allows to to 
publish events to
- * them.
+ * The event handler manages {@link EventListener} instances and allows to 
publish events to them.
  */
 public class TaskEventHandler {
 
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java
index 23096305053..9402a8a3acb 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationHeadTask.java
@@ -74,7 +74,7 @@ import java.util.List;
  * will send a {@link TerminationEvent} to all it's connected tasks, signaling 
them to shutdown.
  *
  * <p>Assumption on the ordering of the outputs: - The first n output gates 
write to channels that
- * go to the tasks of the step function. - The next m output gates to to the 
tasks that consume the
+ * go to the tasks of the step function. - The next m output gates to the 
tasks that consume the
  * final solution. - The last output gate connects to the synchronization task.
  *
  * @param <X> The type of the bulk partial solution / solution set and the 
final output.
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java
index 9fc9c4455cb..608a6a23d7f 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/jobgraph/tasks/TaskOperatorEventGateway.java
@@ -30,8 +30,8 @@ import org.apache.flink.util.SerializedValue;
 import java.util.concurrent.CompletableFuture;
 
 /**
- * Gateway to send an {@link OperatorEvent} or {@link CoordinationRequest} 
from a Task to to the
- * {@link OperatorCoordinator} JobManager side.
+ * Gateway to send an {@link OperatorEvent} or {@link CoordinationRequest} 
from a Task to the {@link
+ * OperatorCoordinator} JobManager side.
  *
  * <p>This is the first step in the chain of sending Operator Events and 
Requests from Operator to
  * Coordinator. Each layer adds further context, so that the inner layers do 
not need to know about
diff --git 
a/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java
 
b/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java
index 8a785226f3c..bfb75a00bb2 100644
--- 
a/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java
+++ 
b/flink-runtime/src/main/java/org/apache/flink/runtime/state/internal/InternalListState.java
@@ -35,7 +35,7 @@ public interface InternalListState<K, N, T>
         extends InternalMergingState<K, N, T, List<T>, Iterable<T>>, 
ListState<T> {
 
     /**
-     * Updates the operator state accessible by {@link #get()} by updating 
existing values to to the
+     * Updates the operator state accessible by {@link #get()} by updating 
existing values to the
      * given list of values. The next time {@link #get()} is called (for the 
same state partition)
      * the returned state will represent the updated list.
      *
diff --git 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java
 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java
index fcd404d5f83..6940707cc9b 100644
--- 
a/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java
+++ 
b/flink-runtime/src/test/java/org/apache/flink/runtime/io/network/partition/consumer/LocalInputChannelTest.java
@@ -401,8 +401,8 @@ public class LocalInputChannelTest {
      * and has not much more general value. If it becomes obsolete at some 
point (future greatness
      * ;)), feel free to remove it.
      *
-     * <p>The fix in the end was to to not acquire the channels lock when 
releasing it and/or not
-     * doing any input gate callbacks while holding the channel's lock. I 
decided to do both.
+     * <p>The fix in the end was to not acquire the channels lock when 
releasing it and/or not doing
+     * any input gate callbacks while holding the channel's lock. I decided to 
do both.
      */
     @Test
     public void testConcurrentReleaseAndRetriggerPartitionRequest() throws 
Exception {
diff --git 
a/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java
 
b/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java
index db58c24f01f..1cc3b6cbd91 100644
--- 
a/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java
+++ 
b/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksDBFullSnapshotResources.java
@@ -90,7 +90,7 @@ public class RocksDBFullSnapshotResources<K> implements 
FullSnapshotResources<K>
         this.keySerializer = keySerializer;
         this.streamCompressionDecorator = streamCompressionDecorator;
 
-        // we need to to this in the constructor, i.e. in the synchronous part 
of the snapshot
+        // we need to do this in the constructor, i.e. in the synchronous part 
of the snapshot
         // TODO: better yet, we can do it outside the constructor
         this.metaData = fillMetaData(metaDataCopy);
     }
diff --git 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java
 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java
index 1616362725d..fbaabb7de9a 100644
--- 
a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java
+++ 
b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/source/ContinuousFileReaderOperator.java
@@ -539,7 +539,7 @@ public class ContinuousFileReaderOperator<OUT, T extends 
TimestampedInputSplit>
             checkpointedState.clear();
 
             throw new Exception(
-                    "Could not add timestamped file input splits to to 
operator "
+                    "Could not add timestamped file input splits to operator "
                             + "state backend of operator "
                             + getOperatorName()
                             + '.',
diff --git 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java
 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java
index c4783660280..d264212c3bd 100644
--- 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java
+++ 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecLegacySink.java
@@ -35,7 +35,7 @@ import javax.annotation.Nullable;
 import java.lang.reflect.Modifier;
 
 /**
- * Batch {@link ExecNode} to to write data into an external sink defined by a 
{@link TableSink}.
+ * Batch {@link ExecNode} to write data into an external sink defined by a 
{@link TableSink}.
  *
  * @param <T> The return type of the {@link TableSink}.
  */
diff --git 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java
 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java
index 25040ee88e6..9a28b1ccf7f 100644
--- 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java
+++ 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/batch/BatchExecSink.java
@@ -44,8 +44,7 @@ import java.util.Collections;
 import java.util.List;
 
 /**
- * Batch {@link ExecNode} to to write data into an external sink defined by a 
{@link
- * DynamicTableSink}.
+ * Batch {@link ExecNode} to write data into an external sink defined by a 
{@link DynamicTableSink}.
  */
 public class BatchExecSink extends CommonExecSink implements 
BatchExecNode<Object> {
     public BatchExecSink(
diff --git 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java
 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java
index 96884df250f..d149f03baa0 100644
--- 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java
+++ 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/common/CommonExecLegacySink.java
@@ -57,7 +57,7 @@ import java.util.Collections;
 import java.util.List;
 
 /**
- * Base {@link ExecNode} to to write data into an external sink defined by a 
{@link TableSink}.
+ * Base {@link ExecNode} to write data into an external sink defined by a 
{@link TableSink}.
  *
  * @param <T> The return type of the {@link TableSink}.
  */
diff --git 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java
 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java
index 04fdde6c11b..82141baac1b 100644
--- 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java
+++ 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecLegacySink.java
@@ -38,7 +38,7 @@ import java.util.List;
 import java.util.stream.Collectors;
 
 /**
- * Stream {@link ExecNode} to to write data into an external sink defined by a 
{@link TableSink}.
+ * Stream {@link ExecNode} to write data into an external sink defined by a 
{@link TableSink}.
  *
  * @param <T> The return type of the {@link TableSink}.
  */
diff --git 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java
 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java
index 79b85a45312..8b6c5614ee0 100644
--- 
a/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java
+++ 
b/flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/stream/StreamExecSink.java
@@ -66,7 +66,7 @@ import java.util.List;
 import java.util.stream.Collectors;
 
 /**
- * Stream {@link ExecNode} to to write data into an external sink defined by a 
{@link
+ * Stream {@link ExecNode} to write data into an external sink defined by a 
{@link
  * DynamicTableSink}.
  */
 @ExecNodeMetadata(
diff --git 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala
 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala
index c9b024bc907..9b4ebe9ad35 100644
--- 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala
+++ 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalLegacySink.scala
@@ -32,7 +32,7 @@ import org.apache.calcite.rel.hint.RelHint
 import java.util
 
 /**
- * Batch physical RelNode to to write data into an external sink defined by a 
[[TableSink]].
+ * Batch physical RelNode to write data into an external sink defined by a 
[[TableSink]].
  *
  * @tparam T
  *   The return type of the [[TableSink]].
diff --git 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala
 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala
index 457b8faee50..8080ec163f3 100644
--- 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala
+++ 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/batch/BatchPhysicalSink.scala
@@ -33,9 +33,7 @@ import org.apache.calcite.rel.hint.RelHint
 
 import java.util
 
-/**
- * Batch physical RelNode to to write data into an external sink defined by a 
[[DynamicTableSink]].
- */
+/** Batch physical RelNode to write data into an external sink defined by a 
[[DynamicTableSink]]. */
 class BatchPhysicalSink(
     cluster: RelOptCluster,
     traitSet: RelTraitSet,
diff --git 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala
 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala
index 1408ed41cde..7a7d95c6dff 100644
--- 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala
+++ 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalLegacySink.scala
@@ -32,7 +32,7 @@ import org.apache.calcite.rel.hint.RelHint
 import java.util
 
 /**
- * Stream physical RelNode to to write data into an external sink defined by a 
[[TableSink]].
+ * Stream physical RelNode to write data into an external sink defined by a 
[[TableSink]].
  *
  * @tparam T
  *   The return type of the [[TableSink]].
diff --git 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala
 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala
index 19b8c600ae7..a058c4b6e68 100644
--- 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala
+++ 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/nodes/physical/stream/StreamPhysicalSink.scala
@@ -36,7 +36,7 @@ import org.apache.calcite.rel.hint.RelHint
 import java.util
 
 /**
- * Stream physical RelNode to to write data into an external sink defined by a 
[[DynamicTableSink]].
+ * Stream physical RelNode to write data into an external sink defined by a 
[[DynamicTableSink]].
  */
 class StreamPhysicalSink(
     cluster: RelOptCluster,
diff --git 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala
 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala
index ffc8c0f54af..fec071307f8 100644
--- 
a/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala
+++ 
b/flink-table/flink-table-planner/src/main/scala/org/apache/flink/table/planner/plan/utils/RelShuttles.scala
@@ -47,7 +47,7 @@ class DefaultRelShuttle extends RelHomogeneousShuttle {
 }
 
 /**
- * Convert all [[QueryOperationCatalogViewTable]]s (including tables in 
[[RexSubQuery]]) to to a
+ * Convert all [[QueryOperationCatalogViewTable]]s (including tables in 
[[RexSubQuery]]) to a
  * relational expression.
  */
 class ExpandTableScanShuttle extends RelShuttleImpl {
diff --git 
a/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java
 
b/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java
index 1e074368fe0..7e4fe0aa16b 100644
--- 
a/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java
+++ 
b/flink-tests/src/test/java/org/apache/flink/test/checkpointing/ProcessingTimeWindowCheckpointingITCase.java
@@ -52,7 +52,7 @@ import static org.junit.Assert.assertTrue;
 import static org.junit.Assert.fail;
 
 /**
- * This test uses a custom non-serializable data type to to ensure that state 
serializability is
+ * This test uses a custom non-serializable data type to ensure that state 
serializability is
  * handled correctly.
  */
 @SuppressWarnings("serial")

Reply via email to