This is an automated email from the ASF dual-hosted git repository. gengliang pushed a commit to branch branch-3.1 in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-3.1 by this push: new ecc3f90 [SPARK-36734][SQL][BUILD][3.1] Upgrade ORC to 1.5.13 ecc3f90 is described below commit ecc3f90578569df4a5548f9555ce2d12c34d3a1c Author: Dongjoon Hyun <dongj...@apache.org> AuthorDate: Thu Sep 16 15:51:27 2021 +0800 [SPARK-36734][SQL][BUILD][3.1] Upgrade ORC to 1.5.13 ### What changes were proposed in this pull request? This PR aims to upgrade Apache ORC from 1.5.12 to 1.5.13 for Apache Spark 3.1.3. ### Why are the changes needed? Apache ORC 1.5.13 is the latest maintenance release in 1.5.x line having the following patches. - https://issues.apache.org/jira/projects/ORC/versions/12349322 ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? Pass the CIs. Closes #33972 from dongjoon-hyun/SPARK-36734. Authored-by: Dongjoon Hyun <dongj...@apache.org> Signed-off-by: Gengliang Wang <gengli...@apache.org> --- dev/deps/spark-deps-hadoop-2.7-hive-2.3 | 6 +++--- dev/deps/spark-deps-hadoop-3.2-hive-2.3 | 6 +++--- pom.xml | 2 +- .../test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala | 4 ++-- 4 files changed, 9 insertions(+), 9 deletions(-) diff --git a/dev/deps/spark-deps-hadoop-2.7-hive-2.3 b/dev/deps/spark-deps-hadoop-2.7-hive-2.3 index 252d823..43f2875 100644 --- a/dev/deps/spark-deps-hadoop-2.7-hive-2.3 +++ b/dev/deps/spark-deps-hadoop-2.7-hive-2.3 @@ -195,9 +195,9 @@ objenesis/2.6//objenesis-2.6.jar okhttp/3.12.12//okhttp-3.12.12.jar okio/1.14.0//okio-1.14.0.jar opencsv/2.3//opencsv-2.3.jar -orc-core/1.5.12//orc-core-1.5.12.jar -orc-mapreduce/1.5.12//orc-mapreduce-1.5.12.jar -orc-shims/1.5.12//orc-shims-1.5.12.jar +orc-core/1.5.13//orc-core-1.5.13.jar +orc-mapreduce/1.5.13//orc-mapreduce-1.5.13.jar +orc-shims/1.5.13//orc-shims-1.5.13.jar oro/2.0.8//oro-2.0.8.jar osgi-resource-locator/1.0.3//osgi-resource-locator-1.0.3.jar paranamer/2.8//paranamer-2.8.jar diff --git a/dev/deps/spark-deps-hadoop-3.2-hive-2.3 b/dev/deps/spark-deps-hadoop-3.2-hive-2.3 index 059f1e1..3fe16b2 100644 --- a/dev/deps/spark-deps-hadoop-3.2-hive-2.3 +++ b/dev/deps/spark-deps-hadoop-3.2-hive-2.3 @@ -209,9 +209,9 @@ okhttp/2.7.5//okhttp-2.7.5.jar okhttp/3.12.12//okhttp-3.12.12.jar okio/1.14.0//okio-1.14.0.jar opencsv/2.3//opencsv-2.3.jar -orc-core/1.5.12//orc-core-1.5.12.jar -orc-mapreduce/1.5.12//orc-mapreduce-1.5.12.jar -orc-shims/1.5.12//orc-shims-1.5.12.jar +orc-core/1.5.13//orc-core-1.5.13.jar +orc-mapreduce/1.5.13//orc-mapreduce-1.5.13.jar +orc-shims/1.5.13//orc-shims-1.5.13.jar oro/2.0.8//oro-2.0.8.jar osgi-resource-locator/1.0.3//osgi-resource-locator-1.0.3.jar paranamer/2.8//paranamer-2.8.jar diff --git a/pom.xml b/pom.xml index 1e8fc7d..bf29905 100644 --- a/pom.xml +++ b/pom.xml @@ -136,7 +136,7 @@ <kafka.version>2.6.0</kafka.version> <derby.version>10.12.1.1</derby.version> <parquet.version>1.10.1</parquet.version> - <orc.version>1.5.12</orc.version> + <orc.version>1.5.13</orc.version> <jetty.version>9.4.40.v20210413</jetty.version> <jakartaservlet.version>4.0.3</jakartaservlet.version> <chill.version>0.9.5</chill.version> diff --git a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala index 876f628..57190a1 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala @@ -695,9 +695,9 @@ class FileBasedDataSourceSuite extends QueryTest test("SPARK-22790,SPARK-27668: spark.sql.sources.compressionFactor takes effect") { Seq(1.0, 0.5).foreach { compressionFactor => withSQLConf(SQLConf.FILE_COMPRESSION_FACTOR.key -> compressionFactor.toString, - SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "250") { + SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "350") { withTempPath { workDir => - // the file size is 486 bytes + // the file size is 504 bytes val workDirPath = workDir.getAbsolutePath val data1 = Seq(100, 200, 300, 400).toDF("count") data1.write.orc(workDirPath + "/data1") --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org