[ 
https://issues.apache.org/jira/browse/SPARK-37994?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17481318#comment-17481318
 ] 

Erik Krogen commented on SPARK-37994:
-------------------------------------

cc [~csun]  – is this related to your changes for using the shaded 
hadoop-client artifact?

> Unable to build spark3.2 with -Dhadoop.version=3.1.4
> ----------------------------------------------------
>
>                 Key: SPARK-37994
>                 URL: https://issues.apache.org/jira/browse/SPARK-37994
>             Project: Spark
>          Issue Type: Bug
>          Components: Build
>    Affects Versions: 3.2.0
>            Reporter: Vu Tan
>            Priority: Minor
>
> I downloaded Spark 3.2 sourcecode from 
> [https://github.com/apache/spark/archive/refs/tags/v3.2.0.zip]
> and try building with the below command 
> {code:java}
> ./dev/make-distribution.sh --name without-hadoop --pip --r --tgz -Psparkr 
> -Phive -Phive-thriftserver -Phadoop-provided -Pyarn -Dhadoop.version=3.1.4 
> -Pkubernetes {code}
> Then it gives the below error 
> {code:java}
> [INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @ 
> spark-core_2.12 ---
> [INFO] Using incremental compilation using Mixed compile order
> [INFO] Compiler bridge file: 
> /Users/JP28431/.sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.15__52.0-1.3.1_20191012T045515.jar
> [INFO] compiler plugin: 
> BasicArtifact(com.github.ghik,silencer-plugin_2.12.15,1.7.6,null)
> [INFO] Compiling 567 Scala sources and 104 Java sources to 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/target/scala-2.12/classes ...
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:38:
>  object io is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:2778:
>  not found: type ArrayWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:2777:
>  not found: type Writable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SSLOptions.scala:24:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SSLOptions.scala:174:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:25:
>  object io is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:26:
>  object security is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala:33:
>  object fs is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala:32:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala:121:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:284:
>  not found: value UserGroupInformation
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:41:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:40:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:39:
>  object mapred is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:37:
>  object fs is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:36:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:348:
>  not found: type Credentials
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:350:
>  not found: value UserGroupInformation
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SecurityManager.scala:402:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:22:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:23:
>  object io is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:24:
>  object io is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:30:
>  not found: type Writable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:38:
>  not found: type ObjectWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SerializableWritable.scala:43:
>  not found: type ObjectWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/internal/io/FileCommitProtocol.scala:20:
>  object fs is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/internal/io/FileCommitProtocol.scala:21:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:210:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:306:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala:26:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala:55:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1188:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1161:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1120:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:925:
>  not found: type TextInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:925:
>  not found: type LongWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:925:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:968:
>  not found: value NewHadoopJob
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:971:
>  not found: value NewFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/NewHadoopRDD.scala:31:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:21:
>  object io is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:39:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:39:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/WholeTextFileInputFormat.scala:25:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/WholeTextFileInputFormat.scala:34:
>  not found: type CombineFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/WholeTextFileRecordReader.scala:21:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/WholeTextFileRecordReader.scala:32:
>  not found: type HConfigurable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:35:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:36:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:20:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/WholeTextFileRDD.scala:37:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:976:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:977:
>  not found: type Text
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1020:
>  not found: value NewHadoopJob
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1023:
>  not found: value NewFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/BinaryFileRDD.scala:22:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/PortableDataStream.scala:28:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/PortableDataStream.scala:39:
>  not found: type CombineFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/BinaryFileRDD.scala:20:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/BinaryFileRDD.scala:34:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1050:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1237:
>  not found: type NewInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala:23:
>  object mapreduce is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/input/FixedLengthBinaryInputFormat.scala:43:
>  not found: type FileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1210:
>  not found: type NewInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1053:
>  not found: type LongWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1053:
>  not found: type BytesWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1055:
>  not found: type LongWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1056:
>  not found: type BytesWritable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1087:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1088:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1096:
>  class FileSystem in package io cannot be accessed in package java.io
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala:34:
>  object mapred is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/deploy/SparkHadoopUtil.scala:131:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:29:
>  object mapred is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:112:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:113:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:99:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:100:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1128:
>  not found: value FileSystem
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/util/SerializableConfiguration.scala:21:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/util/SerializableConfiguration.scala:31:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1132:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1132:
>  not found: value FileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1242:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1247:
>  not found: value FileSystem
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1251:
>  not found: value NewHadoopJob
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1254:
>  not found: value NewFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/NewHadoopRDD.scala:73:
>  not found: type InputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/NewHadoopRDD.scala:27:
>  object conf is not a member of package org.apache.hadoop
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/rdd/NewHadoopRDD.scala:76:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1278:
>  not found: type Configuration
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1277:
>  not found: type NewInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1286:
>  not found: value FileSystem
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1289:
>  not found: type JobConf
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1315:
>  not found: type SequenceFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1375:
>  not found: type SequenceFileInputFormat
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:3045:
>  not found: type Writable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1377:
>  not found: type Writable
> [ERROR] [Error] 
> /Users/JP28431/Downloads/spark-3.2.0-github/core/src/main/scala/org/apache/spark/SparkContext.scala:1378:
>  not found: type Writable
> [INFO] [Info] : Any <: String?
> [INFO] [Info] : false
> [ERROR] 100 errors found
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Reactor Summary for Spark Project Parent POM 3.2.0:
> [INFO]
> [INFO] Spark Project Parent POM ........................... SUCCESS [  4.964 
> s]
> [INFO] Spark Project Tags ................................. SUCCESS [  8.381 
> s]
> [INFO] Spark Project Sketch ............................... SUCCESS [  7.871 
> s]
> [INFO] Spark Project Local DB ............................. SUCCESS [  2.191 
> s]
> [INFO] Spark Project Networking ........................... SUCCESS [  4.879 
> s]
> [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [  2.505 
> s]
> [INFO] Spark Project Unsafe ............................... SUCCESS [  8.310 
> s]
> [INFO] Spark Project Launcher ............................. SUCCESS [  1.990 
> s]
> [INFO] Spark Project Core ................................. FAILURE [ 45.343 
> s]
> [INFO] Spark Project ML Local Library ..................... SKIPPED
> [INFO] Spark Project GraphX ............................... SKIPPED
> [INFO] Spark Project Streaming ............................ SKIPPED
> [INFO] Spark Project Catalyst ............................. SKIPPED
> [INFO] Spark Project SQL .................................. SKIPPED
> [INFO] Spark Project ML Library ........................... SKIPPED
> [INFO] Spark Project Tools ................................ SKIPPED
> [INFO] Spark Project Hive ................................. SKIPPED
> [INFO] Spark Project REPL ................................. SKIPPED
> [INFO] Spark Project YARN Shuffle Service ................. SKIPPED
> [INFO] Spark Project YARN ................................. SKIPPED
> [INFO] Spark Project Kubernetes ........................... SKIPPED
> [INFO] Spark Project Hive Thrift Server ................... SKIPPED
> [INFO] Spark Project Assembly ............................. SKIPPED
> [INFO] Kafka 0.10+ Token Provider for Streaming ........... SKIPPED
> [INFO] Spark Integration for Kafka 0.10 ................... SKIPPED
> [INFO] Kafka 0.10+ Source for Structured Streaming ........ SKIPPED
> [INFO] Spark Project Examples ............................. SKIPPED
> [INFO] Spark Integration for Kafka 0.10 Assembly .......... SKIPPED
> [INFO] Spark Avro ......................................... SKIPPED
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Total time:  01:27 min
> [INFO] Finished at: 2022-01-24T11:07:30+09:00
> [INFO] 
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal 
> net.alchim31.maven:scala-maven-plugin:4.3.0:compile (scala-compile-first) on 
> project spark-core_2.12: Execution scala-compile-first of goal 
> net.alchim31.maven:scala-maven-plugin:4.3.0:compile failed.: CompileFailed -> 
> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/PluginExecutionException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the 
> command
> [ERROR]   mvn <args> -rf :spark-core_2.12 {code}
> is this a expected behavior or should be considered a bug ?



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to