[spark-docker] branch master updated: [SPARK-43372] Use ; instead of && when enable set -ex

2023-05-07 Thread yikun
This is an automated email from the ASF dual-hosted git repository.

yikun pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark-docker.git


The following commit(s) were added to refs/heads/master by this push:
 new 7f9b414  [SPARK-43372] Use ; instead of && when enable set -ex
7f9b414 is described below

commit 7f9b414de48639d69c64acfd81e6792517b86f61
Author: Yikun Jiang 
AuthorDate: Mon May 8 11:19:36 2023 +0800

[SPARK-43372] Use ; instead of && when enable set -ex

### What changes were proposed in this pull request?
- Use ; instead of && when enable set -ex
- ./add-dockerfiles.sh 3.4.0 to apply changes

### Why are the changes needed?
Address DOI comments: `9. using set -ex means you can use ; instead of && 
(really only matters for complex expressions, like the || in the later RUN that 
does use ;)`


https://github.com/docker-library/official-images/pull/13089#issuecomment-1533540388

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
CI passed

Closes #38 from Yikun/SPARK-43372.

Authored-by: Yikun Jiang 
Signed-off-by: Yikun Jiang 
---
 3.4.0/scala2.12-java11-python3-r-ubuntu/Dockerfile | 10 +++
 3.4.0/scala2.12-java11-python3-ubuntu/Dockerfile   |  8 +++---
 3.4.0/scala2.12-java11-r-ubuntu/Dockerfile |  8 +++---
 3.4.0/scala2.12-java11-ubuntu/Dockerfile   | 32 +++---
 Dockerfile.template| 32 +++---
 r-python.template  | 10 +++
 6 files changed, 50 insertions(+), 50 deletions(-)

diff --git a/3.4.0/scala2.12-java11-python3-r-ubuntu/Dockerfile 
b/3.4.0/scala2.12-java11-python3-r-ubuntu/Dockerfile
index 86337c5..12c7a4f 100644
--- a/3.4.0/scala2.12-java11-python3-r-ubuntu/Dockerfile
+++ b/3.4.0/scala2.12-java11-python3-r-ubuntu/Dockerfile
@@ -17,11 +17,11 @@
 ARG BASE_IMAGE=spark:3.4.0-scala2.12-java11-ubuntu
 FROM $BASE_IMAGE
 
-RUN set -ex && \
-apt-get update && \
-apt install -y python3 python3-pip && \
-apt install -y r-base r-base-dev && \
-rm -rf /var/cache/apt/* && \
+RUN set -ex; \
+apt-get update; \
+apt install -y python3 python3-pip; \
+apt install -y r-base r-base-dev; \
+rm -rf /var/cache/apt/*; \
 rm -rf /var/lib/apt/lists/*
 
 ENV R_HOME /usr/lib/R
diff --git a/3.4.0/scala2.12-java11-python3-ubuntu/Dockerfile 
b/3.4.0/scala2.12-java11-python3-ubuntu/Dockerfile
index 540805f..1f0dd1f 100644
--- a/3.4.0/scala2.12-java11-python3-ubuntu/Dockerfile
+++ b/3.4.0/scala2.12-java11-python3-ubuntu/Dockerfile
@@ -17,8 +17,8 @@
 ARG BASE_IMAGE=spark:3.4.0-scala2.12-java11-ubuntu
 FROM $BASE_IMAGE
 
-RUN set -ex && \
-apt-get update && \
-apt install -y python3 python3-pip && \
-rm -rf /var/cache/apt/* && \
+RUN set -ex; \
+apt-get update; \
+apt install -y python3 python3-pip; \
+rm -rf /var/cache/apt/*; \
 rm -rf /var/lib/apt/lists/*
diff --git a/3.4.0/scala2.12-java11-r-ubuntu/Dockerfile 
b/3.4.0/scala2.12-java11-r-ubuntu/Dockerfile
index c65c2ce..53647b2 100644
--- a/3.4.0/scala2.12-java11-r-ubuntu/Dockerfile
+++ b/3.4.0/scala2.12-java11-r-ubuntu/Dockerfile
@@ -17,10 +17,10 @@
 ARG BASE_IMAGE=spark:3.4.0-scala2.12-java11-ubuntu
 FROM $BASE_IMAGE
 
-RUN set -ex && \
-apt-get update && \
-apt install -y r-base r-base-dev && \
-rm -rf /var/cache/apt/* && \
+RUN set -ex; \
+apt-get update; \
+apt install -y r-base r-base-dev; \
+rm -rf /var/cache/apt/*; \
 rm -rf /var/lib/apt/lists/*
 
 ENV R_HOME /usr/lib/R
diff --git a/3.4.0/scala2.12-java11-ubuntu/Dockerfile 
b/3.4.0/scala2.12-java11-ubuntu/Dockerfile
index 21d95d4..11f997f 100644
--- a/3.4.0/scala2.12-java11-ubuntu/Dockerfile
+++ b/3.4.0/scala2.12-java11-ubuntu/Dockerfile
@@ -21,22 +21,22 @@ ARG spark_uid=185
 RUN groupadd --system --gid=${spark_uid} spark && \
 useradd --system --uid=${spark_uid} --gid=spark spark
 
-RUN set -ex && \
-apt-get update && \
-ln -s /lib /lib64 && \
-apt install -y gnupg2 wget bash tini libc6 libpam-modules krb5-user 
libnss3 procps net-tools gosu && \
-mkdir -p /opt/spark && \
-mkdir /opt/spark/python && \
-mkdir -p /opt/spark/examples && \
-mkdir -p /opt/spark/work-dir && \
-chmod g+w /opt/spark/work-dir && \
-touch /opt/spark/RELEASE && \
-chown -R spark:spark /opt/spark && \
-rm /bin/sh && \
-ln -sv /bin/bash /bin/sh && \
-echo "auth required pam_wheel.so use_uid" >> /etc/pam.d/su && \
-chgrp root /etc/passwd && chmod ug+rw /etc/passwd && \
-rm -rf /var/cache/apt/* && \
+RUN set -ex; \
+apt-get update; \
+ln -s /lib /lib64; \
+apt install -y gnupg2 wget bash tini libc6 libpam-modules krb5-user 
libnss3 procps net-tools gosu; \
+mkdir -p /opt/spark; \
+mkdir /opt/spark/python; \
+mkdir -p /opt/spark/examples; \
+mkdir -p /opt/spark/work-dir; \
+chmod 

[spark] branch master updated (8d99e646e98 -> d157b2d9a71)

2023-05-07 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 8d99e646e98 [MINOR][PYTHON] Remove deprecated use of typing.io
 add d157b2d9a71 [MINOR][PYTHON] Fix MyPy linter failure

No new revisions were added by this update.

Summary of changes:
 python/pyspark/broadcast.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [MINOR][PYTHON] Remove deprecated use of typing.io

2023-05-07 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 8d99e646e98 [MINOR][PYTHON] Remove deprecated use of typing.io
8d99e646e98 is described below

commit 8d99e646e983ac1ccb383acf35dcbec745219146
Author: Shantanu <12621235+hauntsani...@users.noreply.github.com>
AuthorDate: Mon May 8 10:47:00 2023 +0900

[MINOR][PYTHON] Remove deprecated use of typing.io

### What changes were proposed in this pull request?
Use `typing.BinaryIO` instead of `typing.io.BinaryIO`. The latter is 
deprecated and had questionable type checker support, see 
https://github.com/python/cpython/issues/92871

### Why are the changes needed?
So Spark is unaffected when `typing.io` is removed in Python 3.13

### Does this PR introduce any user-facing change?
No

### How was this patch tested?
Existing unit tests / every import of this module

Closes #41084 from hauntsaninja/patch-1.

Authored-by: Shantanu <12621235+hauntsani...@users.noreply.github.com>
Signed-off-by: Hyukjin Kwon 
---
 python/pyspark/broadcast.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/python/pyspark/broadcast.py b/python/pyspark/broadcast.py
index a72bf1e059b..605663a2204 100644
--- a/python/pyspark/broadcast.py
+++ b/python/pyspark/broadcast.py
@@ -24,6 +24,7 @@ import pickle
 from typing import (
 overload,
 Any,
+BinaryIO,
 Callable,
 Dict,
 Generic,
@@ -35,7 +36,6 @@ from typing import (
 TYPE_CHECKING,
 Union,
 )
-from typing.io import BinaryIO  # type: ignore[import]
 
 from pyspark.java_gateway import local_connect_and_auth
 from pyspark.serializers import ChunkedStream, pickle_protocol


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (63cb0939679 -> 45a1935fe38)

2023-05-07 Thread kabhwan
This is an automated email from the ASF dual-hosted git repository.

kabhwan pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 63cb0939679 [SPARK-43362][SQL] Special handling of JSON type for MySQL 
connector
 add 45a1935fe38 [SPARK-43364][SS][DOCS] Add docs for RocksDB state store 
memory management

No new revisions were added by this update.

Summary of changes:
 docs/structured-streaming-programming-guide.md | 27 ++
 1 file changed, 27 insertions(+)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-43362][SQL] Special handling of JSON type for MySQL connector

2023-05-07 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 63cb0939679 [SPARK-43362][SQL] Special handling of JSON type for MySQL 
connector
63cb0939679 is described below

commit 63cb09396798da2a16db4bae2b42e0f95bef831b
Author: tianhanhu 
AuthorDate: Mon May 8 10:25:37 2023 +0900

[SPARK-43362][SQL] Special handling of JSON type for MySQL connector

### What changes were proposed in this pull request?

MySQL JSON type is converted into JDBC VARCHAR type with precision of -1 on 
some MariaDB drivers.
When receiving VARCHAR with negative precision, Spark will throw an error.

This PR special cases this scenario by directly converting JSON type into 
StringType in MySQLDialect.

### Why are the changes needed?

Enable reading MySQL tables that has a JSON column.

### Does this PR introduce _any_ user-facing change?

No

### How was this patch tested?

Update existing integration test

Closes #41040 from tianhanhu/SPARK-43362.

Authored-by: tianhanhu 
Signed-off-by: Hyukjin Kwon 
---
 .../scala/org/apache/spark/sql/jdbc/MySQLIntegrationSuite.scala   | 8 +---
 .../src/main/scala/org/apache/spark/sql/jdbc/MySQLDialect.scala   | 4 
 2 files changed, 9 insertions(+), 3 deletions(-)

diff --git 
a/connector/docker-integration-tests/src/test/scala/org/apache/spark/sql/jdbc/MySQLIntegrationSuite.scala
 
b/connector/docker-integration-tests/src/test/scala/org/apache/spark/sql/jdbc/MySQLIntegrationSuite.scala
index c5ca5a72a83..dc3acb66ff1 100644
--- 
a/connector/docker-integration-tests/src/test/scala/org/apache/spark/sql/jdbc/MySQLIntegrationSuite.scala
+++ 
b/connector/docker-integration-tests/src/test/scala/org/apache/spark/sql/jdbc/MySQLIntegrationSuite.scala
@@ -68,10 +68,10 @@ class MySQLIntegrationSuite extends 
DockerJDBCIntegrationSuite {
 
 // TODO: Test locale conversion for strings.
 conn.prepareStatement("CREATE TABLE strings (a CHAR(10), b VARCHAR(10), c 
TINYTEXT, "
-  + "d TEXT, e MEDIUMTEXT, f LONGTEXT, g BINARY(4), h VARBINARY(10), i 
BLOB)"
+  + "d TEXT, e MEDIUMTEXT, f LONGTEXT, g BINARY(4), h VARBINARY(10), i 
BLOB, j JSON)"
 ).executeUpdate()
 conn.prepareStatement("INSERT INTO strings VALUES ('the', 'quick', 
'brown', 'fox', " +
-  "'jumps', 'over', 'the', 'lazy', 'dog')").executeUpdate()
+  "'jumps', 'over', 'the', 'lazy', 'dog', '{\"status\": 
\"merrily\"}')").executeUpdate()
   }
 
   test("Basic test") {
@@ -137,7 +137,7 @@ class MySQLIntegrationSuite extends 
DockerJDBCIntegrationSuite {
 val rows = df.collect()
 assert(rows.length == 1)
 val types = rows(0).toSeq.map(x => x.getClass.toString)
-assert(types.length == 9)
+assert(types.length == 10)
 assert(types(0).equals("class java.lang.String"))
 assert(types(1).equals("class java.lang.String"))
 assert(types(2).equals("class java.lang.String"))
@@ -147,6 +147,7 @@ class MySQLIntegrationSuite extends 
DockerJDBCIntegrationSuite {
 assert(types(6).equals("class [B"))
 assert(types(7).equals("class [B"))
 assert(types(8).equals("class [B"))
+assert(types(9).equals("class java.lang.String"))
 assert(rows(0).getString(0).equals("the".padTo(10, ' ')))
 assert(rows(0).getString(1).equals("quick"))
 assert(rows(0).getString(2).equals("brown"))
@@ -156,6 +157,7 @@ class MySQLIntegrationSuite extends 
DockerJDBCIntegrationSuite {
 assert(java.util.Arrays.equals(rows(0).getAs[Array[Byte]](6), 
Array[Byte](116, 104, 101, 0)))
 assert(java.util.Arrays.equals(rows(0).getAs[Array[Byte]](7), 
Array[Byte](108, 97, 122, 121)))
 assert(java.util.Arrays.equals(rows(0).getAs[Array[Byte]](8), 
Array[Byte](100, 111, 103)))
+assert(rows(0).getString(9).equals("{\"status\": \"merrily\"}"))
   }
 
   test("Basic write test") {
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/jdbc/MySQLDialect.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/jdbc/MySQLDialect.scala
index 5e85ff3ebf6..d6edb67e57e 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/jdbc/MySQLDialect.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/jdbc/MySQLDialect.scala
@@ -98,6 +98,10 @@ private case object MySQLDialect extends JdbcDialect with 
SQLConfHelper {
 } else if ("TINYTEXT".equalsIgnoreCase(typeName)) {
   // TINYTEXT is Types.VARCHAR(63) from mysql jdbc, but keep it AS-IS for 
historical reason
   Some(StringType)
+} else if (sqlType == Types.VARCHAR && typeName.equals("JSON")) {
+  // Some MySQL JDBC drivers converts JSON type into Types.VARCHAR with a 
precision of -1.
+  // Explicitly converts it into StringType here.
+  Some(StringType)
 } else None
   }
 



[spark] branch master updated: [SPARK-43032][SS][CONNECT] Python SQM bug fix

2023-05-07 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 7c5d20b [SPARK-43032][SS][CONNECT] Python SQM bug fix
7c5d20b is described below

commit 7c5d20bb012110d5855e5908cc01658355ed
Author: Wei Liu 
AuthorDate: Mon May 8 10:08:27 2023 +0900

[SPARK-43032][SS][CONNECT] Python SQM bug fix

### What changes were proposed in this pull request?

Some bug fix for streaming ***connect*** python SQM

Note that I also changed ***non-connect***'s StreamingQueryManager `get()` 
API to return an `Optional[StreamingQuery]`.

Before it looks like this when you get a non-exist query:

```
>>> a = spark.streams.get("--0001--0001")
>>> a

>>> a.id
Traceback (most recent call last):
  File "", line 1, in 
  File "/home/wei.liu/oss-spark/python/pyspark/sql/streaming/query.py", 
line 78, in id
return self._jsq.id().toString()
AttributeError: 'NoneType' object has no attribute 'id'
```

But now it looks like:
```
>>> a = spark.streams.get("--0001--0001")
>>> a.id
Traceback (most recent call last):
  File "", line 1, in 
AttributeError: 'NoneType' object has no attribute 'id'
```
The only difference is the return type, which is not typically honored in 
Python... But not very sure if that's a breaking change

### Why are the changes needed?

Bug fix

### Does this PR introduce _any_ user-facing change?

No

### How was this patch tested?

Manually tested. Also verified that it won't throw even without this fix so 
it's not that urgent

Closes #41037 from WweiL/SPARK-43032-python-sqm-fix.

Authored-by: Wei Liu 
Signed-off-by: Hyukjin Kwon 
---
 .../src/main/protobuf/spark/connect/commands.proto |  2 +-
 .../sql/connect/planner/SparkConnectPlanner.scala  | 22 +++--
 python/pyspark/sql/connect/proto/commands_pb2.py   | 36 +++---
 python/pyspark/sql/connect/proto/commands_pb2.pyi  | 16 +-
 python/pyspark/sql/connect/streaming/query.py  | 14 ++---
 python/pyspark/sql/streaming/query.py  |  8 +++--
 6 files changed, 54 insertions(+), 44 deletions(-)

diff --git 
a/connector/connect/common/src/main/protobuf/spark/connect/commands.proto 
b/connector/connect/common/src/main/protobuf/spark/connect/commands.proto
index b929ffa2564..72bc8b5b6ef 100644
--- a/connector/connect/common/src/main/protobuf/spark/connect/commands.proto
+++ b/connector/connect/common/src/main/protobuf/spark/connect/commands.proto
@@ -330,7 +330,7 @@ message StreamingQueryManagerCommand {
 // active() API, returns a list of active queries.
 bool active = 1;
 // get() API, returns the StreamingQuery identified by id.
-string get = 2;
+string get_query = 2;
 // awaitAnyTermination() API, wait until any query terminates or timeout.
 AwaitAnyTerminationCommand await_any_termination = 3;
 // resetTerminated() API.
diff --git 
a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala
 
b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala
index 8c43f982ec1..01f1e890630 100644
--- 
a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala
+++ 
b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala
@@ -2466,16 +2466,18 @@ class SparkConnectPlanner(val session: SparkSession) {
 .toIterable
 .asJava)
 
-  case StreamingQueryManagerCommand.CommandCase.GET =>
-val query = session.streams.get(command.getGet)
-respBuilder.getQueryBuilder
-  .setId(
-StreamingQueryInstanceId
-  .newBuilder()
-  .setId(query.id.toString)
-  .setRunId(query.runId.toString)
-  .build())
-  .setName(SparkConnectService.convertNullString(query.name))
+  case StreamingQueryManagerCommand.CommandCase.GET_QUERY =>
+val query = session.streams.get(command.getGetQuery)
+if (query != null) {
+  respBuilder.getQueryBuilder
+.setId(
+  StreamingQueryInstanceId
+.newBuilder()
+.setId(query.id.toString)
+.setRunId(query.runId.toString)
+.build())
+.setName(SparkConnectService.convertNullString(query.name))
+}
 
   case StreamingQueryManagerCommand.CommandCase.AWAIT_ANY_TERMINATION =>
 if (command.getAwaitAnyTermination.hasTimeoutMs) {
diff --git a/python/pyspark/sql/connect/proto/commands_pb2.py 

[spark] branch branch-3.3 updated (85ff71f9459 -> 9110c05d54c)

2023-05-07 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch branch-3.3
in repository https://gitbox.apache.org/repos/asf/spark.git


from 85ff71f9459 [SPARK-43395][BUILD] Exclude macOS tar extended metadata 
in make-distribution.sh
 add 9110c05d54c [SPARK-37829][SQL][3.3] Dataframe.joinWith outer-join 
should return a null value for unmatched row

No new revisions were added by this update.

Summary of changes:
 .../sql/catalyst/encoders/ExpressionEncoder.scala  | 19 ++---
 .../scala/org/apache/spark/sql/DatasetSuite.scala  | 45 ++
 2 files changed, 58 insertions(+), 6 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated (92ccc60cde5 -> e05e7e0311a)

2023-05-07 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 92ccc60cde5 [SPARK-43181][SQL] Show UI WebURL in `spark-sql` shell
 add e05e7e0311a [MINOR][INFRA] Correct the GitHub PR label for DSTREAM

No new revisions were added by this update.

Summary of changes:
 .github/labeler.yml | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-43181][SQL] Show UI WebURL in `spark-sql` shell

2023-05-07 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 92ccc60cde5 [SPARK-43181][SQL] Show UI WebURL in `spark-sql` shell
92ccc60cde5 is described below

commit 92ccc60cde5c6da2c08b44096f9a34f19c72aa1a
Author: panbingkun 
AuthorDate: Sun May 7 13:59:23 2023 -0700

[SPARK-43181][SQL] Show UI WebURL in `spark-sql` shell

### What changes were proposed in this pull request?
The pr aims to display `the Spark WEB UI address` when spark-sql startup.

### Why are the changes needed?
Promoting user experience.

Like `spark-shell`, it would be great if `spark-sql` show the UI 
information.

```
$ bin/spark-sql
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
23/05/07 13:58:26 WARN NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
23/05/07 13:58:27 WARN HiveConf: HiveConf of name hive.stats.jdbc.timeout 
does not exist
23/05/07 13:58:27 WARN HiveConf: HiveConf of name hive.stats.retries.wait 
does not exist
23/05/07 13:58:28 WARN ObjectStore: Version information not found in 
metastore. hive.metastore.schema.verification is not enabled so recording the 
schema version 2.3.0
23/05/07 13:58:28 WARN ObjectStore: setMetaStoreSchemaVersion called but 
recording version is disabled: version = 2.3.0, comment = Set by MetaStore 
dongjoon127.0.0.1
Spark Web UI available at http://localhost:4040
Spark master: local[*], Application Id: local-1683493106875
spark-sql (default)>
```

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA & Manually test.

Closes #40844 from panbingkun/SPARK-43181.

Authored-by: panbingkun 
Signed-off-by: Dongjoon Hyun 
---
 .../org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala | 3 +++
 1 file changed, 3 insertions(+)

diff --git 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
index c7c905312b7..2302482bb72 100644
--- 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
+++ 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/SparkSQLCLIDriver.scala
@@ -454,6 +454,9 @@ private[hive] class SparkSQLCLIDriver extends CliDriver 
with Logging {
   def printMasterAndAppId(): Unit = {
 val master = SparkSQLEnv.sparkContext.master
 val appId = SparkSQLEnv.sparkContext.applicationId
+SparkSQLEnv.sparkContext.uiWebUrl.foreach {
+  webUrl => console.printInfo(s"Spark Web UI available at $webUrl")
+}
 console.printInfo(s"Spark master: $master, Application Id: $appId")
   }
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-42780][BUILD] Upgrade `Tink` to 1.9.0

2023-05-07 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 96814da9e1a [SPARK-42780][BUILD] Upgrade `Tink` to 1.9.0
96814da9e1a is described below

commit 96814da9e1a2400acc7e31cfc6b3956f7d514031
Author: bjornjorgensen 
AuthorDate: Sun May 7 13:29:37 2023 -0700

[SPARK-42780][BUILD] Upgrade `Tink` to 1.9.0

### What changes were proposed in this pull request?
Upgrade google Tink from 1.7.0 to 1.9.0

[Release note for 
1.8.0](https://github.com/tink-crypto/tink-java/releases/tag/v1.8.0)

[Release note for 
1.9.0](https://github.com/tink-crypto/tink-java/releases/tag/v1.9.0)

### Why are the changes needed?

[SNYK-JAVA-COMGOOGLEPROTOBUF-3040284](https://security.snyk.io/vuln/SNYK-JAVA-COMGOOGLEPROTOBUF-3040284)


[SNYK-JAVA-COMGOOGLEPROTOBUF-3167772](https://security.snyk.io/vuln/SNYK-JAVA-COMGOOGLEPROTOBUF-3167772)

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
This have be benchmarks tested
With
"com.google.crypto.tink" % "tink" % "1.6.1"
(min, avg, max) = (75024163.500, 76331532.832, 77324718.069), stdev = 
652319.870

With
"com.google.crypto.tink" % "tink" % "1.9.0"
(min, avg, max) = (76279051.841, 77512667.749, 78590966.453), stdev = 
632832.384

Almost the same.. Think 1.9.0 is perhaps a bit slower.

Pass GA

Closes #40878 from bjornjorgensen/Tink1.9.0.

Authored-by: bjornjorgensen 
Signed-off-by: Dongjoon Hyun 
---
 dev/deps/spark-deps-hadoop-3-hive-2.3 |  2 +-
 pom.xml   | 12 +++-
 2 files changed, 12 insertions(+), 2 deletions(-)

diff --git a/dev/deps/spark-deps-hadoop-3-hive-2.3 
b/dev/deps/spark-deps-hadoop-3-hive-2.3
index 27df72ecaf8..5174b6b0335 100644
--- a/dev/deps/spark-deps-hadoop-3-hive-2.3
+++ b/dev/deps/spark-deps-hadoop-3-hive-2.3
@@ -245,7 +245,7 @@ stax-api/1.0.1//stax-api-1.0.1.jar
 stream/2.9.6//stream-2.9.6.jar
 super-csv/2.2.0//super-csv-2.2.0.jar
 threeten-extra/1.7.1//threeten-extra-1.7.1.jar
-tink/1.7.0//tink-1.7.0.jar
+tink/1.9.0//tink-1.9.0.jar
 transaction-api/1.1//transaction-api-1.1.jar
 univocity-parsers/2.9.1//univocity-parsers-2.9.1.jar
 wildfly-openssl/1.1.3.Final//wildfly-openssl-1.1.3.Final.jar
diff --git a/pom.xml b/pom.xml
index 96ee3fb5ed9..04260b51c60 100644
--- a/pom.xml
+++ b/pom.xml
@@ -215,7 +215,7 @@
 1.1.0
 1.5.0
 1.60
-1.7.0
+1.9.0
 4.1.89.Final
 

[spark] branch master updated (04ef3d5d0f2 -> 1e090a57f0c)

2023-05-07 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 04ef3d5d0f2 [SPARK-43394][BUILD] Upgrade maven to 3.8.8
 add 1e090a57f0c [SPARK-43347][PYTHON][FOLLOWUP] Change black min 
target-version to py38

No new revisions were added by this update.

Summary of changes:
 dev/pyproject.toml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[spark] branch master updated: [SPARK-43394][BUILD] Upgrade maven to 3.8.8

2023-05-07 Thread srowen
This is an automated email from the ASF dual-hosted git repository.

srowen pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 04ef3d5d0f2 [SPARK-43394][BUILD] Upgrade maven to 3.8.8
04ef3d5d0f2 is described below

commit 04ef3d5d0f2bfebce8dd3b48b9861a2aa5ba1c3a
Author: Cheng Pan 
AuthorDate: Sun May 7 08:24:12 2023 -0500

[SPARK-43394][BUILD] Upgrade maven to 3.8.8

### What changes were proposed in this pull request?

Upgrade Maven from 3.8.7 to 3.8.8.

### Why are the changes needed?

Maven 3.8.8 is the latest patched version of 3.8.x

https://maven.apache.org/docs/3.8.8/release-notes.html

### Does this PR introduce _any_ user-facing change?

No

### How was this patch tested?

Pass GA.

Closes #41073 from pan3793/SPARK-43394.

Authored-by: Cheng Pan 
Signed-off-by: Sean Owen 
---
 dev/appveyor-install-dependencies.ps1 | 2 +-
 docs/building-spark.md| 2 +-
 pom.xml   | 2 +-
 3 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/dev/appveyor-install-dependencies.ps1 
b/dev/appveyor-install-dependencies.ps1
index 7f4f027c820..88090149f5c 100644
--- a/dev/appveyor-install-dependencies.ps1
+++ b/dev/appveyor-install-dependencies.ps1
@@ -81,7 +81,7 @@ if (!(Test-Path $tools)) {
 # == Maven
 # Push-Location $tools
 #
-# $mavenVer = "3.8.7"
+# $mavenVer = "3.8.8"
 # Start-FileDownload 
"https://archive.apache.org/dist/maven/maven-3/$mavenVer/binaries/apache-maven-$mavenVer-bin.zip;
 "maven.zip"
 #
 # # extract
diff --git a/docs/building-spark.md b/docs/building-spark.md
index ba8dddbf6b1..4b8e70655d5 100644
--- a/docs/building-spark.md
+++ b/docs/building-spark.md
@@ -27,7 +27,7 @@ license: |
 ## Apache Maven
 
 The Maven-based build is the build of reference for Apache Spark.
-Building Spark using Maven requires Maven 3.8.7 and Java 8.
+Building Spark using Maven requires Maven 3.8.8 and Java 8/11/17.
 Spark requires Scala 2.12/2.13; support for Scala 2.11 was removed in Spark 
3.0.0.
 
 ### Setting up Maven's Memory Usage
diff --git a/pom.xml b/pom.xml
index c760eaf0cbb..96ee3fb5ed9 100644
--- a/pom.xml
+++ b/pom.xml
@@ -114,7 +114,7 @@
 1.8
 ${java.version}
 ${java.version}
-3.8.7
+3.8.8
 1.6.0
 spark
 2.0.7


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org