This is an automated email from the ASF dual-hosted git repository.

jiayu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/sedona.git


The following commit(s) were added to refs/heads/master by this push:
     new 7abb99ba9 [SEDONA-523] Add pre-commit hook `fix-byte-order-marker` 
(#1301)
7abb99ba9 is described below

commit 7abb99ba943e7911673705b6f187692a1585c2b8
Author: John Bampton <[email protected]>
AuthorDate: Tue Apr 2 09:47:54 2024 +1000

    [SEDONA-523] Add pre-commit hook `fix-byte-order-marker` (#1301)
    
    
https://github.com/pre-commit/pre-commit-hooks?tab=readme-ov-file#fix-byte-order-marker
---
 .pre-commit-config.yaml                | 1 +
 docs/image/geospark-zeppelin-demo.json | 2 +-
 2 files changed, 2 insertions(+), 1 deletion(-)

diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
index 03056953a..8fb3fb2a0 100644
--- a/.pre-commit-config.yaml
+++ b/.pre-commit-config.yaml
@@ -34,6 +34,7 @@ repos:
       - id: detect-private-key
       - id: end-of-file-fixer
         exclude: \.svg$|^docs/image|^spark/common/src/test/resources
+      - id: fix-byte-order-marker
       - id: mixed-line-ending
         exclude: \.csv$
       - id: trailing-whitespace
diff --git a/docs/image/geospark-zeppelin-demo.json 
b/docs/image/geospark-zeppelin-demo.json
index cd71e0781..4f2d53092 100644
--- a/docs/image/geospark-zeppelin-demo.json
+++ b/docs/image/geospark-zeppelin-demo.json
@@ -1 +1 @@
-{"paragraphs":[{"text":"import 
org.datasyslab.geosparksql.utils.GeoSparkSQLRegistrator\nimport 
org.datasyslab.geosparkviz.sql.utils.GeoSparkVizRegistrator\n\n// Zeppelin 
creates and injects sc (SparkContext) and sqlContext (HiveContext or 
SqlContext)\n// So you don't need create them 
manually\n\nGeoSparkSQLRegistrator.registerAll(spark)\nGeoSparkVizRegistrator.registerAll(spark)\n\nvar
 pointdf = spark.read.format(\"csv\").option(\"delimiter\", 
\",\").option(\"header\", \"false\").load(\ [...]
+{"paragraphs":[{"text":"import 
org.datasyslab.geosparksql.utils.GeoSparkSQLRegistrator\nimport 
org.datasyslab.geosparkviz.sql.utils.GeoSparkVizRegistrator\n\n// Zeppelin 
creates and injects sc (SparkContext) and sqlContext (HiveContext or 
SqlContext)\n// So you don't need create them 
manually\n\nGeoSparkSQLRegistrator.registerAll(spark)\nGeoSparkVizRegistrator.registerAll(spark)\n\nvar
 pointdf = spark.read.format(\"csv\").option(\"delimiter\", 
\",\").option(\"header\", \"false\").load(\" [...]

Reply via email to