Hi!

Zinc cache be cleaned by the shutdown i showed you earlier:
./build/zinc-0.3.15/bin/zinc -shutdown

But as I just seen SPARK-34539
<https://issues.apache.org/jira/browse/SPARK-34539>: it is absolutely not
needed anymore as the standalone zinc is not used by the compiler plugin
from v3.0.0.

Regarding make-distribution.sh with hadoop-2.7 I can confirm it is working
on my machine:

$ SPARK_HOME=$PWD ./dev/make-distribution.sh --name custom-spark --pip
 --tgz -Phive -Phive-thriftserver -Pyarn -Phadoop-2.7  -DskipTests
...
[INFO] Reactor Summary for Spark Project Parent POM 3.1.1:
[INFO]
[INFO] Spark Project Parent POM ........................... SUCCESS [
 2.251 s]
[INFO] Spark Project Tags ................................. SUCCESS [
 4.513 s]
[INFO] Spark Project Sketch ............................... SUCCESS [
 4.878 s]
[INFO] Spark Project Local DB ............................. SUCCESS [
 1.259 s]
[INFO] Spark Project Networking ........................... SUCCESS [
 3.173 s]
[INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [
 1.364 s]
[INFO] Spark Project Unsafe ............................... SUCCESS [
 6.672 s]
[INFO] Spark Project Launcher ............................. SUCCESS [
 1.782 s]
[INFO] Spark Project Core ................................. SUCCESS [01:48
min]
[INFO] Spark Project ML Local Library ..................... SUCCESS [
33.861 s]
[INFO] Spark Project GraphX ............................... SUCCESS [
30.114 s]
[INFO] Spark Project Streaming ............................ SUCCESS [
42.267 s]
[INFO] Spark Project Catalyst ............................. SUCCESS [02:15
min]
[INFO] Spark Project SQL .................................. SUCCESS [03:05
min]
[INFO] Spark Project ML Library ........................... SUCCESS [02:16
min]
[INFO] Spark Project Tools ................................ SUCCESS [
 7.109 s]
[INFO] Spark Project Hive ................................. SUCCESS [01:20
min]
[INFO] Spark Project REPL ................................. SUCCESS [
20.758 s]
[INFO] Spark Project YARN Shuffle Service ................. SUCCESS [
10.377 s]
[INFO] Spark Project YARN ................................. SUCCESS [
47.571 s]
[INFO] Spark Project Hive Thrift Server ................... SUCCESS [
36.327 s]
[INFO] Spark Project Assembly ............................. SUCCESS [
 4.618 s]
[INFO] Kafka 0.10+ Token Provider for Streaming ........... SUCCESS [
18.895 s]
[INFO] Spark Integration for Kafka 0.10 ................... SUCCESS [
28.380 s]
[INFO] Kafka 0.10+ Source for Structured Streaming ........ SUCCESS [
 02:00 h]
[INFO] Spark Project Examples ............................. SUCCESS [26:28
min]
[INFO] Spark Integration for Kafka 0.10 Assembly .......... SUCCESS [
 4.884 s]
[INFO] Spark Avro ......................................... SUCCESS [
35.016 s]
[INFO]
------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO]
------------------------------------------------------------------------
[INFO] Total time:  02:43 h
[INFO] Finished at: 2021-03-11T08:12:02+01:00
[INFO]
------------------------------------------------------------------------
...
Creating tar archive
removing 'pyspark-3.1.1' (and everything under it)
+ popd
+ '[' false == true ']'
+ echo 'Skipping building R source package'
Skipping building R source package
+ mkdir /Users/attilazsoltpiros/git/attilapiros/spark/dist/conf
+ cp
/Users/attilazsoltpiros/git/attilapiros/spark/conf/fairscheduler.xml.template
/Users/attilazsoltpiros/git/attilapiros/spark/conf/log4j.properties.template
/Users/attilazsoltpiros/git/attilapiros/spark/conf/metrics.properties.template
/Users/attilazsoltpiros/git/attilapiros/spark/conf/spark-defaults.conf.template
/Users/attilazsoltpiros/git/attilapiros/spark/conf/spark-env.sh.template
/Users/attilazsoltpiros/git/attilapiros/spark/conf/workers.template
/Users/attilazsoltpiros/git/attilapiros/spark/dist/conf
+ cp /Users/attilazsoltpiros/git/attilapiros/spark/README.md
/Users/attilazsoltpiros/git/attilapiros/spark/dist
+ cp -r /Users/attilazsoltpiros/git/attilapiros/spark/bin
/Users/attilazsoltpiros/git/attilapiros/spark/dist
+ cp -r /Users/attilazsoltpiros/git/attilapiros/spark/python
/Users/attilazsoltpiros/git/attilapiros/spark/dist
+ '[' true == true ']'
+ rm -f
/Users/attilazsoltpiros/git/attilapiros/spark/dist/python/dist/pyspark-3.1.1.tar.gz
/Users/attilazsoltpiros/git/attilapiros/spark/dist/python/dist/pyspark-3.2.0.dev0.tar.gz
+ cp -r /Users/attilazsoltpiros/git/attilapiros/spark/sbin
/Users/attilazsoltpiros/git/attilapiros/spark/dist
+ '[' -d /Users/attilazsoltpiros/git/attilapiros/spark/R/lib/SparkR ']'
+ mkdir -p /Users/attilazsoltpiros/git/attilapiros/spark/dist/R/lib
+ cp -r /Users/attilazsoltpiros/git/attilapiros/spark/R/lib/SparkR
/Users/attilazsoltpiros/git/attilapiros/spark/dist/R/lib
+ cp /Users/attilazsoltpiros/git/attilapiros/spark/R/lib/sparkr.zip
/Users/attilazsoltpiros/git/attilapiros/spark/dist/R/lib
+ '[' true == true ']'
+ TARDIR_NAME=spark-3.1.1-bin-custom-spark
+
TARDIR=/Users/attilazsoltpiros/git/attilapiros/spark/spark-3.1.1-bin-custom-spark
+ rm -rf
/Users/attilazsoltpiros/git/attilapiros/spark/spark-3.1.1-bin-custom-spark
+ cp -r /Users/attilazsoltpiros/git/attilapiros/spark/dist
/Users/attilazsoltpiros/git/attilapiros/spark/spark-3.1.1-bin-custom-spark
+ tar czf spark-3.1.1-bin-custom-spark.tgz -C
/Users/attilazsoltpiros/git/attilapiros/spark spark-3.1.1-bin-custom-spark
+ rm -rf
/Users/attilazsoltpiros/git/attilapiros/spark/spark-3.1.1-bin-custom-spark

Best Regards,
Attila



On Thu, Mar 11, 2021 at 5:11 AM jiahong li <monkeyboy....@gmail.com> wrote:

> BTW, how to  clear zinc/nailgun cache, thanks
>
> jiahong li <monkeyboy....@gmail.com> 于2021年3月11日周四 下午12:04写道:
>
>> Maybe it is my environment cause
>>
>> jiahong li <monkeyboy....@gmail.com> 于2021年3月11日周四 上午11:14写道:
>>
>>> it not the cause,when i set -Phadoop-2.7 instead of
>>> -Dhadoop.version=2.6.0-cdh5.13.1, the same errors come out.
>>>
>>> Attila Zsolt Piros <piros.attila.zs...@gmail.com> 于2021年3月10日周三
>>> 下午8:56写道:
>>>
>>>> I see, this must be because of hadoop version you are selecting by
>>>> using "-Dhadoop.version=2.6.0-cdh5.13.1".
>>>> Spark 3.1.1 only support hadoop-2.7 and hadoop-3.2, at least these two
>>>> can be given via profiles:  -Phadoop-2.7  and -Phadoop-3.2 (the default).
>>>>
>>>>
>>>> On Wed, Mar 10, 2021 at 12:26 PM jiahong li <monkeyboy....@gmail.com>
>>>> wrote:
>>>>
>>>>> i use ./build/mvn to compile ,and after execute command 
>>>>> :./build/zinc-0.3.15/bin/zinc
>>>>> -shutdown
>>>>> and execute command like this: /dev/make-distribution.sh --name
>>>>> custom-spark --pip  --tgz -Phive -Phive-thriftserver -Pyarn
>>>>> -Dhadoop.version=2.6.0-cdh5.13.1 -DskipTests
>>>>> same error appear.
>>>>> and execute command: ps -ef |grep zinc, there is nothing containe zinc
>>>>>
>>>>> Attila Zsolt Piros <piros.attila.zs...@gmail.com> 于2021年3月10日周三
>>>>> 下午6:55写道:
>>>>>
>>>>>> hi!
>>>>>>
>>>>>> Are you compiling Spark itself?
>>>>>> Do you use "./build/mvn" from the project root?
>>>>>> If you compiled an other version of Spark before and there the scala
>>>>>> version was different then zinc/nailgun could cached the old classes 
>>>>>> which
>>>>>> can cause similar troubles.
>>>>>> In that case this could help:
>>>>>>
>>>>>> ./build/zinc-0.3.15/bin/zinc -shutdown
>>>>>>
>>>>>> Best Regards,
>>>>>> Attila
>>>>>>
>>>>>> On Wed, Mar 10, 2021 at 11:27 AM jiahong li <monkeyboy....@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> hi, everybody, when i compile spark 3.1.1 from tag v3.1.1 ,encounter
>>>>>>> error like this:
>>>>>>>
>>>>>>> INFO] --- scala-maven-plugin:4.3.0:compile (scala-compile-first) @
>>>>>>> spark-core_2.12 ---
>>>>>>> [INFO] Using incremental compilation using Mixed compile order
>>>>>>> [INFO] Compiler bridge file:
>>>>>>> .sbt/1.0/zinc/org.scala-sbt/org.scala-sbt-compiler-bridge_2.12-1.3.1-bin_2.12.10__52.0-1.3.1_20191012T045515.jar
>>>>>>> [INFO] compiler plugin:
>>>>>>> BasicArtifact(com.github.ghik,silencer-plugin_2.12.10,1.6.0,null)
>>>>>>> [INFO] Compiling 560 Scala sources and 99 Java sources to
>>>>>>> git/spark/core/target/scala-2.12/classes ...
>>>>>>> [ERROR] [Error]
>>>>>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>>>>>> type mismatch;
>>>>>>>  found   : K where type K
>>>>>>>  required: String
>>>>>>> [ERROR] [Error]
>>>>>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>>>>>> value map is not a member of V
>>>>>>> [ERROR] [Error]
>>>>>>> git/spark/core/src/main/scala/org/apache/spark/ui/HttpSecurityFilter.scala:107:
>>>>>>> missing argument list for method stripXSS in class XssSafeRequest
>>>>>>> Unapplied methods are only converted to functions when a function
>>>>>>> type is expected.
>>>>>>> You can make this conversion explicit by writing `stripXSS _` or
>>>>>>> `stripXSS(_)` instead of `stripXSS`.
>>>>>>> [ERROR] [Error]
>>>>>>> git/spark/core/src/main/scala/org/apache/spark/ui/PagedTable.scala:307:
>>>>>>> value startsWith is not a member of K
>>>>>>> [ERROR] [Error]
>>>>>>> git/spark/core/src/main/scala/org/apache/spark/util/Utils.scala:580: 
>>>>>>> value
>>>>>>> toLowerCase is not a member of object org.apache.hadoop.util.StringUtils
>>>>>>> [ERROR] 5 errors found
>>>>>>>
>>>>>>> anybody encounter error like this?
>>>>>>>
>>>>>>>
>>>>>>

Reply via email to