This is an automated email from the ASF dual-hosted git repository.
kirs pushed a commit to branch auto-pick-55675-branch-3.1
in repository https://gitbox.apache.org/repos/asf/doris.git
The following commit(s) were added to refs/heads/auto-pick-55675-branch-3.1 by
this push:
new 6b556b6b6cf check style
6b556b6b6cf is described below
commit 6b556b6b6cfa0c13c9af99ec4f9ea6c8eb769ba1
Author: Calvin Kirs <[email protected]>
AuthorDate: Tue Sep 9 21:16:20 2025 +0800
check style
---
.../org/apache/doris/datasource/CatalogMgrTest.java | 20 ++++++++++----------
.../datasource/property/PropertyConverterTest.java | 6 +++---
2 files changed, 13 insertions(+), 13 deletions(-)
diff --git
a/fe/fe-core/src/test/java/org/apache/doris/datasource/CatalogMgrTest.java
b/fe/fe-core/src/test/java/org/apache/doris/datasource/CatalogMgrTest.java
index 2ac7b8682b5..2a7422cc55d 100644
--- a/fe/fe-core/src/test/java/org/apache/doris/datasource/CatalogMgrTest.java
+++ b/fe/fe-core/src/test/java/org/apache/doris/datasource/CatalogMgrTest.java
@@ -125,7 +125,7 @@ public class CatalogMgrTest extends TestWithFeService {
env.getCatalogMgr().createCatalog(hiveCatalog2);
CreateCatalogStmt iceBergCatalog = (CreateCatalogStmt)
parseAndAnalyzeStmt(
- "create catalog iceberg properties('type' = 'hms',
'iceberg.hive.metastore.uris' = 'thrift://192.168.0.1:9083');",
+ "create catalog iceberg properties('type' = 'hms',
'hive.metastore.uris' = 'thrift://192.168.0.1:9083');",
rootCtx);
env.getCatalogMgr().createCatalog(iceBergCatalog);
@@ -573,7 +573,7 @@ public class CatalogMgrTest extends TestWithFeService {
+
"='org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider'\n"
+ ");";
CreateCatalogStmt createStmt1 = (CreateCatalogStmt)
parseAndAnalyzeStmt(createCatalogSql);
- ExceptionChecker.expectThrowsWithMsg(DdlException.class, "Missing
dfs.ha.namenodes.your-nameservice property",
+ ExceptionChecker.expectThrowsWithMsg(IllegalArgumentException.class,
"Missing property: dfs.ha.namenodes.your-nameservice",
() -> mgr.createCatalog(createStmt1));
createCatalogSql = "CREATE CATALOG bad_hive2 PROPERTIES (\n"
@@ -587,8 +587,8 @@ public class CatalogMgrTest extends TestWithFeService {
+
"='org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider'\n"
+ ");";
CreateCatalogStmt createStmt2 = (CreateCatalogStmt)
parseAndAnalyzeStmt(createCatalogSql);
- ExceptionChecker.expectThrowsWithMsg(DdlException.class,
- "Missing dfs.namenode.rpc-address.your-nameservice.nn1
property",
+ ExceptionChecker.expectThrowsWithMsg(IllegalArgumentException.class,
+ "Missing property:
dfs.namenode.rpc-address.your-nameservice.nn1 (expected format: host:port)",
() -> mgr.createCatalog(createStmt2));
createCatalogSql = "CREATE CATALOG good_hive PROPERTIES (\n"
@@ -601,8 +601,8 @@ public class CatalogMgrTest extends TestWithFeService {
+ "
'dfs.namenode.rpc-address.your-nameservice.nn2'='172.21.0.3:4007'\n"
+ ");";
CreateCatalogStmt createStmt3 = (CreateCatalogStmt)
parseAndAnalyzeStmt(createCatalogSql);
- ExceptionChecker.expectThrowsWithMsg(DdlException.class,
- "Missing dfs.client.failover.proxy.provider.your-nameservice
property",
+ ExceptionChecker.expectThrowsWithMsg(IllegalArgumentException.class,
+ "Missing property:
dfs.client.failover.proxy.provider.your-nameservice",
() -> mgr.createCatalog(createStmt3));
createCatalogSql = "CREATE CATALOG bad_jdbc PROPERTIES (\n"
@@ -690,8 +690,8 @@ public class CatalogMgrTest extends TestWithFeService {
+ ");";
AlterCatalogPropertyStmt alterCatalogPropertyStmt1 =
(AlterCatalogPropertyStmt) parseAndAnalyzeStmt(
alterCatalogSql);
- ExceptionChecker.expectThrowsWithMsg(DdlException.class,
- "Missing dfs.ha.namenodes.HANN property",
+ ExceptionChecker.expectThrowsWithMsg(IllegalArgumentException.class,
+ "Missing property: dfs.ha.namenodes.HANN",
() -> mgr.alterCatalogProps(alterCatalogPropertyStmt1));
alterCatalogSql = "ALTER CATALOG test_hive1 SET PROPERTIES (\n"
@@ -706,8 +706,8 @@ public class CatalogMgrTest extends TestWithFeService {
+ ");";
AlterCatalogPropertyStmt alterCatalogPropertyStmt2 =
(AlterCatalogPropertyStmt) parseAndAnalyzeStmt(
alterCatalogSql);
- ExceptionChecker.expectThrowsWithMsg(DdlException.class,
- "Missing dfs.client.failover.proxy.provider.HANN property",
+ ExceptionChecker.expectThrowsWithMsg(IllegalArgumentException.class,
+ "Missing property: dfs.client.failover.proxy.provider.HANN",
() -> mgr.alterCatalogProps(alterCatalogPropertyStmt2));
alterCatalogSql = "ALTER CATALOG test_hive1 SET PROPERTIES (\n"
diff --git
a/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
b/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
index 6e9686481d0..cdf8a63d4ea 100644
---
a/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
+++
b/fe/fe-core/src/test/java/org/apache/doris/datasource/property/PropertyConverterTest.java
@@ -259,8 +259,8 @@ public class PropertyConverterTest extends
TestWithFeService {
String catalogName = "hms_oss_hdfs";
CreateCatalogStmt analyzedStmt = createStmt(query1);
HMSExternalCatalog catalog = createAndGetCatalog(analyzedStmt,
catalogName);
- Map<String, String> hdProps =
catalog.getCatalogProperty().getHadoopProperties();
- Assertions.assertEquals("cn-beijing.oss-dls.aliyuncs.com",
hdProps.get("fs.oss.endpoint"));
+ Map<String, String> hdProps =
catalog.getCatalogProperty().getBackendStorageProperties();
+
Assertions.assertTrue(hdProps.get("fs.oss.endpoint").endsWith("aliyuncs.com"));
}
@Disabled
@@ -756,7 +756,7 @@ public class PropertyConverterTest extends
TestWithFeService {
String createIceGlue = "CREATE CATALOG iceglue PROPERTIES (\n"
+ " \"type\"=\"iceberg\",\n"
+ " \"iceberg.catalog.type\" = \"glue\",\n"
- + " \"glue.endpoint\" =
\"https://glue.us-east-1.amazonaws.com/\",\n"
+ + " \"glue.endpoint\" =
\"https://glue.us-east-1.amazonaws.com\",\n"
+ " \"glue.access_key\" = \"ak123\",\n"
+ " \"glue.secret_key\" = \"sk123\"\n"
+ ");";
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]