This is an automated email from the ASF dual-hosted git repository.
lzljs3620320 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/paimon.git
The following commit(s) were added to refs/heads/master by this push:
new 093c72e7c5 [test][spark] Fix the unstable testHiveCatalogOptions
(#5738)
093c72e7c5 is described below
commit 093c72e7c592341c8c8603b883153c8d683cf4b1
Author: Zouxxyy <[email protected]>
AuthorDate: Thu Jun 12 18:12:32 2025 +0800
[test][spark] Fix the unstable testHiveCatalogOptions (#5738)
---
.../paimon/spark/SparkCatalogWithHiveTest.java | 1 +
.../spark/SparkGenericCatalogWithHiveTest.java | 30 ++++++++++------------
.../catalog/functions/BucketFunctionTest.java | 6 +++++
3 files changed, 20 insertions(+), 17 deletions(-)
diff --git
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkCatalogWithHiveTest.java
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkCatalogWithHiveTest.java
index 9f383a54c2..c7c07b1aef 100644
---
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkCatalogWithHiveTest.java
+++
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkCatalogWithHiveTest.java
@@ -109,6 +109,7 @@ public class SparkCatalogWithHiveTest {
.collect(Collectors.toList()))
.containsExactlyInAnyOrder("[1,1,1]", "[2,2,2]");
}
+ spark.stop();
}
@Test
diff --git
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkGenericCatalogWithHiveTest.java
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkGenericCatalogWithHiveTest.java
index 3bb013648e..9e13dd7be5 100644
---
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkGenericCatalogWithHiveTest.java
+++
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/SparkGenericCatalogWithHiveTest.java
@@ -60,12 +60,15 @@ public class SparkGenericCatalogWithHiveTest {
.config(
"spark.sql.catalog.spark_catalog",
SparkGenericCatalog.class.getName())
+ .config("spark.sql.catalog.paimon.warehouse",
warehousePath.toString())
+ .config("spark.sql.catalog.paimon",
SparkCatalog.class.getName())
.config(
"spark.sql.extensions",
"org.apache.paimon.spark.extensions.PaimonSparkSessionExtensions")
.master("local[2]")
.getOrCreate();
+ spark.sql("USE spark_catalog");
spark.sql("CREATE DATABASE my_db");
spark.sql("USE my_db");
spark.sql(
@@ -80,29 +83,22 @@ public class SparkGenericCatalogWithHiveTest {
.map(s -> s.get(1))
.map(Object::toString))
.containsExactlyInAnyOrder("t1");
- spark.close();
- // secondly, we close catalog with hive metastore, and start a
filesystem metastore to check
- // the result.
- SparkSession spark2 =
- SparkSession.builder()
- .config("spark.sql.catalog.paimon.warehouse",
warehousePath.toString())
- .config("spark.sql.catalogImplementation", "in-memory")
- .config("spark.sql.catalog.paimon",
SparkCatalog.class.getName())
- .config(
- "spark.sql.extensions",
-
"org.apache.paimon.spark.extensions.PaimonSparkSessionExtensions")
- .master("local[2]")
- .getOrCreate();
- spark2.sql("USE paimon");
- spark2.sql("USE my_db");
- assertThat(spark2.sql("SHOW
NAMESPACES").collectAsList().stream().map(Object::toString))
+ // secondly, use filesystem metastore to check the result.
+ spark.sql("USE paimon");
+ spark.sql("USE my_db");
+ assertThat(spark.sql("SHOW
NAMESPACES").collectAsList().stream().map(Object::toString))
.containsExactlyInAnyOrder("[default]", "[my_db]");
assertThat(
- spark2.sql("SHOW TABLES").collectAsList().stream()
+ spark.sql("SHOW TABLES").collectAsList().stream()
.map(s -> s.get(1))
.map(Object::toString))
.containsExactlyInAnyOrder("t1");
+
+ spark.sql("USE spark_catalog");
+ spark.sql("DROP TABLE my_db.t1");
+ spark.sql("DROP DATABASE my_db");
+ spark.close();
}
@Test
diff --git
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/catalog/functions/BucketFunctionTest.java
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/catalog/functions/BucketFunctionTest.java
index cf02654063..b8fbcdae42 100644
---
a/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/catalog/functions/BucketFunctionTest.java
+++
b/paimon-spark/paimon-spark-ut/src/test/java/org/apache/paimon/spark/catalog/functions/BucketFunctionTest.java
@@ -56,6 +56,7 @@ import
org.apache.paimon.shade.guava30.com.google.common.collect.ImmutableMap;
import org.apache.spark.sql.SparkSession;
import org.assertj.core.api.Assertions;
+import org.junit.jupiter.api.AfterAll;
import org.junit.jupiter.api.AfterEach;
import org.junit.jupiter.api.BeforeAll;
import org.junit.jupiter.api.Test;
@@ -190,6 +191,11 @@ public class BucketFunctionTest {
spark.sql(String.format("DROP TABLE IF EXISTS %s", TABLE_NAME));
}
+ @AfterAll
+ public static void tearDown() {
+ spark.stop();
+ }
+
public static void setupTable(String... bucketColumns) {
String commitUser = UUID.randomUUID().toString();
try {