This is an automated email from the ASF dual-hosted git repository.
jiayu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/sedona.git
The following commit(s) were added to refs/heads/master by this push:
new 7abb99ba9 [SEDONA-523] Add pre-commit hook `fix-byte-order-marker`
(#1301)
7abb99ba9 is described below
commit 7abb99ba943e7911673705b6f187692a1585c2b8
Author: John Bampton <[email protected]>
AuthorDate: Tue Apr 2 09:47:54 2024 +1000
[SEDONA-523] Add pre-commit hook `fix-byte-order-marker` (#1301)
https://github.com/pre-commit/pre-commit-hooks?tab=readme-ov-file#fix-byte-order-marker
---
.pre-commit-config.yaml | 1 +
docs/image/geospark-zeppelin-demo.json | 2 +-
2 files changed, 2 insertions(+), 1 deletion(-)
diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
index 03056953a..8fb3fb2a0 100644
--- a/.pre-commit-config.yaml
+++ b/.pre-commit-config.yaml
@@ -34,6 +34,7 @@ repos:
- id: detect-private-key
- id: end-of-file-fixer
exclude: \.svg$|^docs/image|^spark/common/src/test/resources
+ - id: fix-byte-order-marker
- id: mixed-line-ending
exclude: \.csv$
- id: trailing-whitespace
diff --git a/docs/image/geospark-zeppelin-demo.json
b/docs/image/geospark-zeppelin-demo.json
index cd71e0781..4f2d53092 100644
--- a/docs/image/geospark-zeppelin-demo.json
+++ b/docs/image/geospark-zeppelin-demo.json
@@ -1 +1 @@
-{"paragraphs":[{"text":"import
org.datasyslab.geosparksql.utils.GeoSparkSQLRegistrator\nimport
org.datasyslab.geosparkviz.sql.utils.GeoSparkVizRegistrator\n\n// Zeppelin
creates and injects sc (SparkContext) and sqlContext (HiveContext or
SqlContext)\n// So you don't need create them
manually\n\nGeoSparkSQLRegistrator.registerAll(spark)\nGeoSparkVizRegistrator.registerAll(spark)\n\nvar
pointdf = spark.read.format(\"csv\").option(\"delimiter\",
\",\").option(\"header\", \"false\").load(\ [...]
+{"paragraphs":[{"text":"import
org.datasyslab.geosparksql.utils.GeoSparkSQLRegistrator\nimport
org.datasyslab.geosparkviz.sql.utils.GeoSparkVizRegistrator\n\n// Zeppelin
creates and injects sc (SparkContext) and sqlContext (HiveContext or
SqlContext)\n// So you don't need create them
manually\n\nGeoSparkSQLRegistrator.registerAll(spark)\nGeoSparkVizRegistrator.registerAll(spark)\n\nvar
pointdf = spark.read.format(\"csv\").option(\"delimiter\",
\",\").option(\"header\", \"false\").load(\" [...]