[
https://issues.apache.org/jira/browse/LIVY-1010?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Gyorgy Gal updated LIVY-1010:
-----------------------------
Fix Version/s: (was: 0.8.0)
> Add support for Spark 3.5.4
> ---------------------------
>
> Key: LIVY-1010
> URL: https://issues.apache.org/jira/browse/LIVY-1010
> Project: Livy
> Issue Type: Improvement
> Affects Versions: 0.8.0
> Reporter: Mnr Bsf
> Assignee: Gyorgy Gal
> Priority: Major
>
> It will be good to keep the Apache Livy project up to date with the latest
> Spark version (3.5.4).
> Tried Apache Livy 0.8 for Spark 3.5.4 (with Java 11 and 17), and did not work
> as expected like Apache Livy 0.8 and Spark version 3.0.0.
> Java 17, I was not able to start an Apache Livy session
> {code:java}
> Exception in thread "main" java.util.concurrent.ExecutionException:
> javax.security.sasl.SaslException: Client closed before SASL negotiation
> finished {code}
> Java 11, I was able to start an Apache Livy session but running any sample
> code was throwing this exception
> {code:java}
> 'JavaPackage' object is not callable
> Traceback (most recent call last):
> File "/opt/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 1443,
> in createDataFrame
> return self._create_dataframe(
> File "/opt/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 1485,
> in _create_dataframe
> rdd, struct = self._createFromLocal(map(prepare, data), schema)
> File "/opt/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 1093,
> in _createFromLocal
> struct = self._inferSchemaFromList(data, names=schema)
> File "/opt/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 954,
> in _inferSchemaFromList
> prefer_timestamp_ntz = is_timestamp_ntz_preferred()
> File "/opt/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 153, in
> is_timestamp_ntz_preferred
> return jvm is not None and jvm.PythonSQLUtils.isTimestampNTZPreferred()
> TypeError: 'JavaPackage' object is not callable {code}
> The only way to make Spark 3.5.4 working with Java 11 is to issue some
> java_imports when a session starts
> {code:java}
> from py4j.java_gateway import java_import
> java_import(spark._sc._jvm, "org.apache.spark.SparkConf")
> java_import(spark._sc._jvm, "org.apache.spark.api.java.*")
> java_import(spark._sc._jvm, "org.apache.spark.api.python.*")
> java_import(spark._sc._jvm, "org.apache.spark.ml.python.*")
> java_import(spark._sc._jvm, "org.apache.spark.mllib.api.python.*")
> java_import(spark._sc._jvm, "org.apache.spark.resource.*")
> java_import(spark._sc._jvm, "org.apache.spark.sql.*")
> java_import(spark._sc._jvm, "org.apache.spark.sql.api.python.*")
> java_import(spark._sc._jvm, "org.apache.spark.sql.hive.*") {code}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)