zhangfengcdt opened a new pull request, #1455:
URL: https://github.com/apache/sedona/pull/1455
- Added default Spark configurations in `TestBaseScala` trait.
- Provided a method `defaultSparkConfig` for default configurations.
- Allowed subclasses to override `sparkConfig` to add or modify Spark
configurations.
- Updated initialization of Spark session and Spark context to use the
provided configurations.
- Included comments to explain how to override configurations in subclasses.
- Ensured that default configurations are preserved if not overridden by
subclasses.
## Did you read the Contributor Guide?
- Yes, I have read the [Contributor
Rules](https://sedona.apache.org/latest-snapshot/community/rule/) and
[Contributor Development
Guide](https://sedona.apache.org/latest-snapshot/community/develop/)
## Is this PR related to a JIRA ticket?
- Yes, the URL of the associated JIRA ticket is
https://issues.apache.org/jira/browse/SEDONA-568. The PR name follows the
format `[SEDONA-XXX] my subject`.
## What changes were proposed in this PR?
- Added default Spark configurations in `TestBaseScala` trait.
- Provided a method `defaultSparkConfig` for default configurations.
- Allowed subclasses to override `sparkConfig` to add or modify Spark
configurations.
- Updated initialization of Spark session and Spark context to use the
provided configurations.
- Included comments to explain how to override configurations in subclasses.
- Ensured that default configurations are preserved if not overridden by
subclasses.
After this change, user can either use
- Default Configurations: Provided in defaultSparkConfig.
- Overriding Configurations: Subclasses can override sparkConfig and merge
additional configurations with the default ones using the ++ operator, for
example:
```
class CustomTest extends TestBaseScala {
// Override sparkConfig to provide additional configurations
override def sparkConfig: Map[String, String] = defaultSparkConfig ++ Map(
"spark.sql.warehouse.dir" -> "/custom/warehouse/location",
"spark.some.config.option" -> "some-value"
)
describe("CustomTest") {
it("should use custom Spark configuration") {
assert(sparkSession.conf.get("spark.some.config.option") ==
"some-value")
}
it("should access Spark context") {
assert(sc != null)
}
}
}
```
## How was this patch tested?
All existing unit tests should all pass without any impact.
## Did this PR include necessary documentation updates?
- No, this PR does not affect any public API so no need to change the
documentation.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: dev-unsubscr...@sedona.apache.org
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org