This is an automated email from the ASF dual-hosted git repository.

czweng pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/flink-table-store.git

commit e367d98740fbccf6b58730adc79bda338ffd60c5
Author: tsreaper <tsreape...@gmail.com>
AuthorDate: Thu Dec 8 17:04:05 2022 +0800

    [hotfix] Set checkpointing.randomization to false to avoid random unaligned 
checkpoints in tests and also delete useless TableEnvironmentTestUtils
---
 .../table/store/connector/CatalogITCaseBase.java   |  8 +---
 .../ChangelogWithKeyFileStoreTableITCase.java      |  5 +-
 .../CompositePkAndMultiPartitionedTableITCase.java | 19 ++++----
 .../store/connector/FileStoreTableITCase.java      |  8 +---
 .../table/store/connector/LookupJoinITCase.java    |  4 +-
 .../table/store/connector/MappingTableITCase.java  |  4 +-
 .../store/connector/ReadWriteTableITCase.java      | 17 ++++---
 .../store/connector/ReadWriteTableTestBase.java    |  5 +-
 .../store/connector/StreamingWarehouseITCase.java  |  4 +-
 .../store/connector/TableEnvironmentTestUtils.java | 54 ----------------------
 .../table/store/connector/TableStoreTestBase.java  |  3 +-
 .../store/connector/sink/SinkSavepointITCase.java  |  5 +-
 pom.xml                                            |  2 +-
 13 files changed, 33 insertions(+), 105 deletions(-)

diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CatalogITCaseBase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CatalogITCaseBase.java
index 17d066cd..8893464d 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CatalogITCaseBase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CatalogITCaseBase.java
@@ -49,9 +49,7 @@ public abstract class CatalogITCaseBase extends 
AbstractTestBase {
 
     @Before
     public void before() throws IOException {
-        tEnv =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inBatchMode().build());
+        tEnv = 
TableEnvironment.create(EnvironmentSettings.newInstance().inBatchMode().build());
         String catalog = "TABLE_STORE";
         tEnv.executeSql(
                 String.format(
@@ -59,9 +57,7 @@ public abstract class CatalogITCaseBase extends 
AbstractTestBase {
                         catalog, TEMPORARY_FOLDER.newFolder().toURI()));
         tEnv.useCatalog(catalog);
 
-        sEnv =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inStreamingMode().build());
+        sEnv = 
TableEnvironment.create(EnvironmentSettings.newInstance().inStreamingMode().build());
         sEnv.getConfig().getConfiguration().set(CHECKPOINTING_INTERVAL, 
Duration.ofMillis(100));
         sEnv.registerCatalog(catalog, tEnv.getCatalog(catalog).get());
         sEnv.useCatalog(catalog);
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ChangelogWithKeyFileStoreTableITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ChangelogWithKeyFileStoreTableITCase.java
index 42c34e40..607d703f 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ChangelogWithKeyFileStoreTableITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ChangelogWithKeyFileStoreTableITCase.java
@@ -99,13 +99,12 @@ public class ChangelogWithKeyFileStoreTableITCase extends 
TestBaseUtils {
     }
 
     private TableEnvironment createBatchTableEnvironment() {
-        return TableEnvironmentTestUtils.create(
-                EnvironmentSettings.newInstance().inBatchMode().build());
+        return 
TableEnvironment.create(EnvironmentSettings.newInstance().inBatchMode().build());
     }
 
     private TableEnvironment createStreamingTableEnvironment() {
         TableEnvironment sEnv =
-                TableEnvironmentTestUtils.create(
+                TableEnvironment.create(
                         
EnvironmentSettings.newInstance().inStreamingMode().build());
         // set checkpoint interval to a random number to emulate different 
speed of commit
         sEnv.getConfig()
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CompositePkAndMultiPartitionedTableITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CompositePkAndMultiPartitionedTableITCase.java
index f0886d05..a0a6b3f6 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CompositePkAndMultiPartitionedTableITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/CompositePkAndMultiPartitionedTableITCase.java
@@ -87,7 +87,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -273,7 +273,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -402,7 +402,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -495,7 +495,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -598,7 +598,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -737,8 +737,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // batch read to check data refresh
         final StreamTableEnvironment batchTableEnv =
-                TableEnvironmentTestUtils.create(
-                        buildBatchEnv(), EnvironmentSettings.inBatchMode());
+                StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         registerTable(batchTableEnv, managedTable);
         collectAndCheck(
                 batchTableEnv,
@@ -843,8 +842,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // batch read to check data refresh
         final StreamTableEnvironment batchTableEnv =
-                TableEnvironmentTestUtils.create(
-                        buildBatchEnv(), EnvironmentSettings.inBatchMode());
+                StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         registerTable(batchTableEnv, managedTable);
         collectAndCheck(
                 batchTableEnv,
@@ -925,8 +923,7 @@ public class CompositePkAndMultiPartitionedTableITCase 
extends ReadWriteTableTes
 
         // batch read to check data refresh
         final StreamTableEnvironment batchTableEnv =
-                TableEnvironmentTestUtils.create(
-                        buildBatchEnv(), EnvironmentSettings.inBatchMode());
+                StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         registerTable(batchTableEnv, managedTable);
         collectAndCheck(
                 batchTableEnv,
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/FileStoreTableITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/FileStoreTableITCase.java
index d5ff515d..3bb66a62 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/FileStoreTableITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/FileStoreTableITCase.java
@@ -65,12 +65,8 @@ public abstract class FileStoreTableITCase extends 
AbstractTestBase {
 
     @Before
     public void before() throws IOException {
-        bEnv =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inBatchMode().build());
-        sEnv =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inStreamingMode().build());
+        bEnv = 
TableEnvironment.create(EnvironmentSettings.newInstance().inBatchMode().build());
+        sEnv = 
TableEnvironment.create(EnvironmentSettings.newInstance().inStreamingMode().build());
         sEnv.getConfig().getConfiguration().set(CHECKPOINTING_INTERVAL, 
Duration.ofMillis(100));
         path = TEMPORARY_FOLDER.newFolder().toURI().toString();
         prepareConfiguration(bEnv, path);
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/LookupJoinITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/LookupJoinITCase.java
index a0f502b9..54c7c34f 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/LookupJoinITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/LookupJoinITCase.java
@@ -43,9 +43,7 @@ public class LookupJoinITCase extends AbstractTestBase {
 
     @Before
     public void before() throws Exception {
-        env =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inStreamingMode().build());
+        env = 
TableEnvironment.create(EnvironmentSettings.newInstance().inStreamingMode().build());
         env.getConfig().getConfiguration().set(CHECKPOINTING_INTERVAL, 
Duration.ofMillis(100));
         env.getConfig()
                 .getConfiguration()
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/MappingTableITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/MappingTableITCase.java
index d68ace6d..7f909390 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/MappingTableITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/MappingTableITCase.java
@@ -44,9 +44,7 @@ public class MappingTableITCase extends AbstractTestBase {
 
     @Before
     public void before() throws IOException {
-        tEnv =
-                TableEnvironmentTestUtils.create(
-                        
EnvironmentSettings.newInstance().inBatchMode().build());
+        tEnv = 
TableEnvironment.create(EnvironmentSettings.newInstance().inBatchMode().build());
         path = TEMPORARY_FOLDER.newFolder().toURI().toString();
     }
 
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableITCase.java
index b4a56c39..0cc8fc07 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableITCase.java
@@ -92,7 +92,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         BlockingIterator<Row, Row> streamIter =
                 collectAndCheck(
@@ -211,7 +211,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
 
         // test streaming read
         final StreamTableEnvironment streamTableEnv =
-                TableEnvironmentTestUtils.create(buildStreamEnv());
+                StreamTableEnvironment.create(buildStreamEnv());
         registerTable(streamTableEnv, managedTable);
         collectAndCheck(
                         streamTableEnv,
@@ -428,8 +428,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
 
         // batch read to check data refresh
         final StreamTableEnvironment batchTableEnv =
-                TableEnvironmentTestUtils.create(
-                        buildBatchEnv(), EnvironmentSettings.inBatchMode());
+                StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         registerTable(batchTableEnv, managedTable);
         collectAndCheck(
                 batchTableEnv,
@@ -1208,7 +1207,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
     @Test
     public void testQueryContainsDefaultFieldName() throws Exception {
         rootPath = TEMPORARY_FOLDER.newFolder().getPath();
-        tEnv = TableEnvironmentTestUtils.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
+        tEnv = StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         String id = registerData(Collections.singletonList(changelogRow("+I", 
1, "abc")));
         tEnv.executeSql(
                 String.format(
@@ -1234,7 +1233,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
     @Test
     public void testLike() throws Exception {
         rootPath = TEMPORARY_FOLDER.newFolder().getPath();
-        tEnv = TableEnvironmentTestUtils.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
+        tEnv = StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         List<Row> input =
                 Arrays.asList(
                         changelogRow("+I", 1, "test_1"),
@@ -1335,7 +1334,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
     @Test
     public void testIn() throws Exception {
         rootPath = TEMPORARY_FOLDER.newFolder().getPath();
-        tEnv = TableEnvironmentTestUtils.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
+        tEnv = StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         List<Row> input =
                 Arrays.asList(
                         changelogRow("+I", 1, "aaa"),
@@ -1416,7 +1415,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
     @Test
     public void testChangeBucketNumber() throws Exception {
         rootPath = TEMPORARY_FOLDER.newFolder().getPath();
-        tEnv = TableEnvironmentTestUtils.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
+        tEnv = StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         tEnv.executeSql(
                 String.format(
                         "CREATE TABLE IF NOT EXISTS rates (\n"
@@ -1484,7 +1483,7 @@ public class ReadWriteTableITCase extends 
ReadWriteTableTestBase {
     public void testStreamingInsertOverwrite() throws Exception {
         rootPath = TEMPORARY_FOLDER.newFolder().getPath();
         tEnv =
-                TableEnvironmentTestUtils.create(
+                StreamTableEnvironment.create(
                         buildStreamEnv(), 
EnvironmentSettings.inStreamingMode());
         tEnv.executeSql(
                 String.format(
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableTestBase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableTestBase.java
index 41867850..db9170f7 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableTestBase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/ReadWriteTableTestBase.java
@@ -313,7 +313,7 @@ public class ReadWriteTableTestBase extends 
KafkaTableTestBase {
             env = buildBatchEnv();
             builder.inBatchMode();
         }
-        tEnv = TableEnvironmentTestUtils.create(env, builder.build());
+        tEnv = StreamTableEnvironment.create(env, builder.build());
         tEnv.executeSql(helperTableDdl);
 
         String managedTableDdl;
@@ -406,8 +406,7 @@ public class ReadWriteTableTestBase extends 
KafkaTableTestBase {
 
     protected void prepareEnvAndOverwrite(String managedTable, String query) 
throws Exception {
         final StreamTableEnvironment batchEnv =
-                TableEnvironmentTestUtils.create(
-                        buildBatchEnv(), EnvironmentSettings.inBatchMode());
+                StreamTableEnvironment.create(buildBatchEnv(), 
EnvironmentSettings.inBatchMode());
         registerTable(batchEnv, managedTable);
         batchEnv.executeSql(query).await();
     }
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/StreamingWarehouseITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/StreamingWarehouseITCase.java
index 7516df9d..faa8eef7 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/StreamingWarehouseITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/StreamingWarehouseITCase.java
@@ -36,9 +36,9 @@ import static org.assertj.core.api.Assertions.assertThat;
 public class StreamingWarehouseITCase extends ReadWriteTableTestBase {
 
     private final StreamTableEnvironment streamTableEnv =
-            TableEnvironmentTestUtils.create(buildStreamEnv(1));
+            StreamTableEnvironment.create(buildStreamEnv(1));
     private final StreamTableEnvironment batchTableEnv =
-            TableEnvironmentTestUtils.create(buildBatchEnv(1), 
EnvironmentSettings.inBatchMode());
+            StreamTableEnvironment.create(buildBatchEnv(1), 
EnvironmentSettings.inBatchMode());
 
     @Test
     public void testUserStory() throws Exception {
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableEnvironmentTestUtils.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableEnvironmentTestUtils.java
deleted file mode 100644
index f8588ec2..00000000
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableEnvironmentTestUtils.java
+++ /dev/null
@@ -1,54 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.flink.table.store.connector;
-
-import 
org.apache.flink.streaming.api.environment.ExecutionCheckpointingOptions;
-import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
-import org.apache.flink.table.api.EnvironmentSettings;
-import org.apache.flink.table.api.TableEnvironment;
-import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
-
-/** Utility class for creating {@link TableEnvironment} in tests. */
-public class TableEnvironmentTestUtils {
-
-    public static TableEnvironment create(EnvironmentSettings settings) {
-        TableEnvironment tEnv = TableEnvironment.create(settings);
-        disableUnalignedCheckpoint(tEnv);
-        return tEnv;
-    }
-
-    public static StreamTableEnvironment create(StreamExecutionEnvironment 
env) {
-        StreamTableEnvironment tEnv = StreamTableEnvironment.create(env);
-        disableUnalignedCheckpoint(tEnv);
-        return tEnv;
-    }
-
-    public static StreamTableEnvironment create(
-            StreamExecutionEnvironment env, EnvironmentSettings settings) {
-        StreamTableEnvironment tEnv = StreamTableEnvironment.create(env, 
settings);
-        disableUnalignedCheckpoint(tEnv);
-        return tEnv;
-    }
-
-    private static void disableUnalignedCheckpoint(TableEnvironment tEnv) {
-        tEnv.getConfig()
-                .getConfiguration()
-                .set(ExecutionCheckpointingOptions.ENABLE_UNALIGNED, false);
-    }
-}
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableStoreTestBase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableStoreTestBase.java
index 0d5868fa..f8728183 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableStoreTestBase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/TableStoreTestBase.java
@@ -22,6 +22,7 @@ import org.apache.flink.api.common.RuntimeExecutionMode;
 import org.apache.flink.configuration.Configuration;
 import org.apache.flink.table.api.EnvironmentSettings;
 import org.apache.flink.table.api.Schema;
+import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
 import org.apache.flink.table.api.internal.TableEnvironmentImpl;
 import org.apache.flink.table.catalog.CatalogTable;
 import org.apache.flink.table.catalog.Column;
@@ -90,7 +91,7 @@ public abstract class TableStoreTestBase extends 
KafkaTableTestBase {
             env.enableCheckpointing(100);
             builder.inStreamingMode();
         }
-        tEnv = TableEnvironmentTestUtils.create(env, builder.build());
+        tEnv = StreamTableEnvironment.create(env, builder.build());
         ((TableEnvironmentImpl) tEnv)
                 .getCatalogManager()
                 .registerCatalog(
diff --git 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/sink/SinkSavepointITCase.java
 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/sink/SinkSavepointITCase.java
index ca4614da..1aa4546f 100644
--- 
a/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/sink/SinkSavepointITCase.java
+++ 
b/flink-table-store-connector/src/test/java/org/apache/flink/table/store/connector/sink/SinkSavepointITCase.java
@@ -33,7 +33,6 @@ import org.apache.flink.table.api.EnvironmentSettings;
 import org.apache.flink.table.api.TableEnvironment;
 import org.apache.flink.table.api.bridge.java.StreamTableEnvironment;
 import org.apache.flink.table.api.config.ExecutionConfigOptions;
-import org.apache.flink.table.store.connector.TableEnvironmentTestUtils;
 import org.apache.flink.table.store.file.utils.FailingAtomicRenameFileSystem;
 import org.apache.flink.test.util.AbstractTestBase;
 import org.apache.flink.types.Row;
@@ -134,7 +133,7 @@ public class SinkSavepointITCase extends AbstractTestBase {
 
         EnvironmentSettings settings = 
EnvironmentSettings.newInstance().inStreamingMode().build();
         StreamExecutionEnvironment env = 
StreamExecutionEnvironment.getExecutionEnvironment(conf);
-        StreamTableEnvironment tEnv = TableEnvironmentTestUtils.create(env, 
settings);
+        StreamTableEnvironment tEnv = StreamTableEnvironment.create(env, 
settings);
         tEnv.getConfig()
                 .getConfiguration()
                 .set(ExecutionCheckpointingOptions.CHECKPOINTING_INTERVAL, 
Duration.ofMillis(500));
@@ -206,7 +205,7 @@ public class SinkSavepointITCase extends AbstractTestBase {
 
     private void checkRecoverFromSavepointResult(String failingPath) throws 
Exception {
         EnvironmentSettings settings = 
EnvironmentSettings.newInstance().inBatchMode().build();
-        TableEnvironment tEnv = TableEnvironmentTestUtils.create(settings);
+        TableEnvironment tEnv = TableEnvironment.create(settings);
         // no failure should occur when checking for answer
         FailingAtomicRenameFileSystem.reset(failingName, 0, 1);
 
diff --git a/pom.xml b/pom.xml
index f067f1be..40ee8dac 100644
--- a/pom.xml
+++ b/pom.xml
@@ -444,7 +444,7 @@ under the License.
                     <trimStackTrace>false</trimStackTrace>
                     <systemPropertyVariables>
                         <forkNumber>0${surefire.forkNumber}</forkNumber>
-                        
<checkpointing.randomization>true</checkpointing.randomization>
+                        
<checkpointing.randomization>false</checkpointing.randomization>
                         <project.basedir>${project.basedir}</project.basedir>
                         <!--suppress MavenModelInspection -->
                         
<test.randomization.seed>${test.randomization.seed}</test.randomization.seed>

Reply via email to