This is an automated email from the ASF dual-hosted git repository.
lzljs3620320 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/paimon.git
The following commit(s) were added to refs/heads/master by this push:
new daaca523cf [test] Add format end to Spark migrate file test
daaca523cf is described below
commit daaca523cf97de33c188ecbf9830d3ec8e4957bc
Author: JingsongLi <[email protected]>
AuthorDate: Mon Mar 17 23:10:55 2025 +0800
[test] Add format end to Spark migrate file test
---
.../spark/procedure/MigrateFileProcedureTest.scala | 84 ++++++++++++----------
1 file changed, 47 insertions(+), 37 deletions(-)
diff --git
a/paimon-spark/paimon-spark-ut/src/test/scala/org/apache/paimon/spark/procedure/MigrateFileProcedureTest.scala
b/paimon-spark/paimon-spark-ut/src/test/scala/org/apache/paimon/spark/procedure/MigrateFileProcedureTest.scala
index 492350a0f7..0121fd6c39 100644
---
a/paimon-spark/paimon-spark-ut/src/test/scala/org/apache/paimon/spark/procedure/MigrateFileProcedureTest.scala
+++
b/paimon-spark/paimon-spark-ut/src/test/scala/org/apache/paimon/spark/procedure/MigrateFileProcedureTest.scala
@@ -31,29 +31,31 @@ class MigrateFileProcedureTest extends PaimonHiveTestBase {
Seq("parquet", "orc", "avro").foreach(
format => {
test(s"Paimon migrate file procedure: migrate $format non-partitioned
table") {
- withTable(s"hive_tbl1$random", s"paimon_tbl1$random") {
+ withTable(s"hive_tbl1$random$format", s"paimon_tbl1$random$format") {
// create hive table
spark.sql(s"""
- |CREATE TABLE hive_tbl1$random (id STRING, name STRING,
pt STRING)
+ |CREATE TABLE hive_tbl1$random$format (id STRING, name
STRING, pt STRING)
|USING $format
|""".stripMargin)
- spark.sql(s"INSERT INTO hive_tbl1$random VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
+ spark.sql(
+ s"INSERT INTO hive_tbl1$random$format VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
// create paimon table
spark.sql(s"""
- |CREATE TABLE paimon_tbl1$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE paimon_tbl1$random$format (id STRING,
name STRING, pt STRING)
|USING PAIMON
|TBLPROPERTIES ('file.format'='$format', 'bucket'='-1')
|""".stripMargin)
- spark.sql(s"INSERT INTO paimon_tbl1$random VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
+ spark.sql(
+ s"INSERT INTO paimon_tbl1$random$format VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
spark.sql(
- s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl1$random', target_table =>
'$hiveDbName.paimon_tbl1$random')")
+ s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl1$random$format', target_table =>
'$hiveDbName.paimon_tbl1$random$format')")
checkAnswer(
- spark.sql(s"SELECT * FROM paimon_tbl1$random ORDER BY id"),
+ spark.sql(s"SELECT * FROM paimon_tbl1$random$format ORDER BY id"),
Row("1", "a", "p1") :: Row("2", "b", "p2") :: Row("3", "c", "p1")
:: Row(
"4",
"d",
@@ -66,29 +68,31 @@ class MigrateFileProcedureTest extends PaimonHiveTestBase {
format => {
test(
s"Paimon migrate file procedure: migrate $format non-partitioned table
with parallelism") {
- withTable(s"hive_tbl_02$random", s"paimon_tbl_02$random") {
+ withTable(s"hive_tbl_02$random$format",
s"paimon_tbl_02$random$format") {
// create hive table
spark.sql(s"""
- |CREATE TABLE hive_tbl_02$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE hive_tbl_02$random$format (id STRING,
name STRING, pt STRING)
|USING $format
|""".stripMargin)
- spark.sql(s"INSERT INTO hive_tbl_02$random VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
+ spark.sql(
+ s"INSERT INTO hive_tbl_02$random$format VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
// create paimon table
spark.sql(s"""
- |CREATE TABLE paimon_tbl_02$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE paimon_tbl_02$random$format (id STRING,
name STRING, pt STRING)
|USING PAIMON
|TBLPROPERTIES ('file.format'='$format', 'bucket'='-1')
|""".stripMargin)
- spark.sql(s"INSERT INTO paimon_tbl_02$random VALUES ('3', 'c',
'p1'), ('4', 'd', 'p2')")
+ spark.sql(
+ s"INSERT INTO paimon_tbl_02$random$format VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
spark.sql(
- s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl_02$random', target_table =>
'$hiveDbName.paimon_tbl_02$random', parallelism => 6)")
+ s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl_02$random$format', target_table =>
'$hiveDbName.paimon_tbl_02$random$format', parallelism => 6)")
checkAnswer(
- spark.sql(s"SELECT * FROM paimon_tbl_02$random ORDER BY id"),
+ spark.sql(s"SELECT * FROM paimon_tbl_02$random$format ORDER BY
id"),
Row("1", "a", "p1") :: Row("2", "b", "p2") :: Row("3", "c", "p1")
:: Row(
"4",
"d",
@@ -101,31 +105,33 @@ class MigrateFileProcedureTest extends PaimonHiveTestBase
{
format => {
test(
s"Paimon migrate file procedure: migrate $format non-partitioned table
with delete source table") {
- withTable(s"hive_tbl3$random", s"paimon_tbl3$random") {
+ withTable(s"hive_tbl3$random$format", s"paimon_tbl3$random$format") {
// create hive table
spark.sql(s"""
- |CREATE TABLE hive_tbl3$random (id STRING, name STRING,
pt STRING)
+ |CREATE TABLE hive_tbl3$random$format (id STRING, name
STRING, pt STRING)
|USING $format
|""".stripMargin)
- spark.sql(s"INSERT INTO hive_tbl3$random VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
+ spark.sql(
+ s"INSERT INTO hive_tbl3$random$format VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
// create paimon table
spark.sql(s"""
- |CREATE TABLE paimon_tbl3$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE paimon_tbl3$random$format (id STRING,
name STRING, pt STRING)
|USING PAIMON
|TBLPROPERTIES ('file.format'='$format', 'bucket'='-1')
|""".stripMargin)
- spark.sql(s"INSERT INTO paimon_tbl3$random VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
+ spark.sql(
+ s"INSERT INTO paimon_tbl3$random$format VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
spark.sql(
- s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl3$random', target_table =>
'$hiveDbName.paimon_tbl3$random', delete_origin => false)")
+ s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl3$random$format', target_table =>
'$hiveDbName.paimon_tbl3$random$format', delete_origin => false)")
- checkAnswer(spark.sql(s"SELECT * FROM hive_tbl3$random ORDER BY
id"), Nil)
+ checkAnswer(spark.sql(s"SELECT * FROM hive_tbl3$random$format ORDER
BY id"), Nil)
checkAnswer(
- spark.sql(s"SELECT * FROM paimon_tbl3$random ORDER BY id"),
+ spark.sql(s"SELECT * FROM paimon_tbl3$random$format ORDER BY id"),
Row("1", "a", "p1") :: Row("2", "b", "p2") :: Row("3", "c", "p1")
:: Row(
"4",
"d",
@@ -137,31 +143,33 @@ class MigrateFileProcedureTest extends PaimonHiveTestBase
{
Seq("parquet", "orc", "avro").foreach(
format => {
test(s"Paimon migrate file procedure: migrate $format partitioned
table") {
- withTable(s"hive_tbl4$random", s"paimon_tbl4$random") {
+ withTable(s"hive_tbl4$random$format", s"paimon_tbl4$random$format") {
// create hive table
spark.sql(s"""
- |CREATE TABLE hive_tbl4$random (id STRING, name STRING,
pt STRING)
+ |CREATE TABLE hive_tbl4$random$format (id STRING, name
STRING, pt STRING)
|USING $format
|PARTITIONED BY (pt)
|""".stripMargin)
- spark.sql(s"INSERT INTO hive_tbl4$random VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
+ spark.sql(
+ s"INSERT INTO hive_tbl4$random$format VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
// create paimon table
spark.sql(s"""
- |CREATE TABLE paimon_tbl4$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE paimon_tbl4$random$format (id STRING,
name STRING, pt STRING)
|USING PAIMON
|TBLPROPERTIES ('file.format'='$format', 'bucket'='-1')
|PARTITIONED BY (pt)
|""".stripMargin)
- spark.sql(s"INSERT INTO paimon_tbl4$random VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
+ spark.sql(
+ s"INSERT INTO paimon_tbl4$random$format VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
spark.sql(
- s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl4$random', target_table =>
'$hiveDbName.paimon_tbl4$random')")
+ s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl4$random$format', target_table =>
'$hiveDbName.paimon_tbl4$random$format')")
checkAnswer(
- spark.sql(s"SELECT * FROM paimon_tbl4$random ORDER BY id"),
+ spark.sql(s"SELECT * FROM paimon_tbl4$random$format ORDER BY id"),
Row("1", "a", "p1") :: Row("2", "b", "p2") :: Row("3", "c", "p1")
:: Row(
"4",
"d",
@@ -174,37 +182,39 @@ class MigrateFileProcedureTest extends PaimonHiveTestBase
{
format => {
test(
s"Paimon migrate file procedure: migrate $format partitioned table
with delete source table") {
- withTable(s"hive_tbl5$random", s"paimon_tbl5$random") {
+ withTable(s"hive_tbl5$random$format", s"paimon_tbl5$random$format") {
// create hive table
spark.sql(s"""
- |CREATE TABLE hive_tbl5$random (id STRING, name STRING,
pt STRING)
+ |CREATE TABLE hive_tbl5$random$format (id STRING, name
STRING, pt STRING)
|USING $format
|PARTITIONED BY (pt)
|""".stripMargin)
- spark.sql(s"INSERT INTO hive_tbl5$random VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
+ spark.sql(
+ s"INSERT INTO hive_tbl5$random$format VALUES ('1', 'a', 'p1'),
('2', 'b', 'p2')")
// create paimon table
spark.sql(s"""
- |CREATE TABLE paimon_tbl5$random (id STRING, name
STRING, pt STRING)
+ |CREATE TABLE paimon_tbl5$random$format (id STRING,
name STRING, pt STRING)
|USING PAIMON
|TBLPROPERTIES ('file.format'='$format', 'bucket'='-1')
|PARTITIONED BY (pt)
|""".stripMargin)
- spark.sql(s"INSERT INTO paimon_tbl5$random VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
+ spark.sql(
+ s"INSERT INTO paimon_tbl5$random$format VALUES ('3', 'c', 'p1'),
('4', 'd', 'p2')")
spark.sql(
- s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl5$random', target_table =>
'$hiveDbName.paimon_tbl5$random', delete_origin => false)")
+ s"CALL sys.migrate_table(source_type => 'hive', table =>
'$hiveDbName.hive_tbl5$random$format', target_table =>
'$hiveDbName.paimon_tbl5$random$format', delete_origin => false)")
checkAnswer(
- spark.sql(s"SELECT * FROM paimon_tbl5$random ORDER BY id"),
+ spark.sql(s"SELECT * FROM paimon_tbl5$random$format ORDER BY id"),
Row("1", "a", "p1") :: Row("2", "b", "p2") :: Row("3", "c", "p1")
:: Row(
"4",
"d",
"p2") :: Nil)
- checkAnswer(spark.sql(s"SELECT * FROM hive_tbl5$random ORDER BY
id"), Nil)
+ checkAnswer(spark.sql(s"SELECT * FROM hive_tbl5$random$format ORDER
BY id"), Nil)
}
}
})