Re: Object in compiler mirror not found - maven build
Or you just have zinc running but in a bad state. `zinc -shutdown` should kill it off and let you try again. On Sun, Nov 26, 2017 at 2:12 PM, Sean Owenwrote: > I'm not seeing that on OS X or Linux. It sounds a bit like you have an old > version of zinc or scala or something installed. > > On Sun, Nov 26, 2017 at 3:55 PM Tomasz Dudek < > megatrontomaszdu...@gmail.com> wrote: > >> Hello everyone, >> >> I would love to help develop Apache Spark. I have run into a (very >> basic?) issue which holds me in that mission. >> >> I followed the `how to contribute` guide, however running ./build/mvn >> -DskipTests clean package fails with: >> >> [INFO] Using zinc server for incremental compilation >> [info] 'compiler-interface' not yet compiled for Scala 2.11.8. >> Compiling... >> error: scala.reflect.internal.MissingRequirementError: object >> java.lang.Object in compiler mirror not found. >> at scala.reflect.internal.MissingRequirementError$.signal( >> MissingRequirementError.scala:17) >> at scala.reflect.internal.MissingRequirementError$.notFound( >> MissingRequirementError.scala:18) >> at scala.reflect.internal.Mirrors$RootsBase. >> getModuleOrClass(Mirrors.scala:53) >> >> is it perhaps compability issue? Versions I use are as follows: >> >> ➜ spark git:(master) ✗ ./build/mvn --version >> Using `mvn` from path: /Users/tdudek/Programming/ >> spark/build/apache-maven-3.3.9/bin/mvn >> Apache Maven 3.3.9 (bb52d8502b132ec0a5a3f4c09453c07478323dc5; >> 2015-11-10T17:41:47+01:00) >> Maven home: /Users/tdudek/Programming/spark/build/apache-maven-3.3.9 >> Java version: 1.8.0_152, vendor: Oracle Corporation >> Java home: /Library/Java/JavaVirtualMachines/jdk1.8.0_ >> 152.jdk/Contents/Home/jre >> Default locale: en_PL, platform encoding: US-ASCII >> OS name: "mac os x", version: "10.13.1", arch: "x86_64", family: "mac" >> >> I just lost few hours mindlessly trying to make it work. I hate to waste >> other peoples' time and I'm REALLY ashamed at my question, but I think I am >> missing something fundamental. >> >> Cheers, >> Tomasz >> >
Re: Object in compiler mirror not found - maven build
I'm not seeing that on OS X or Linux. It sounds a bit like you have an old version of zinc or scala or something installed. On Sun, Nov 26, 2017 at 3:55 PM Tomasz Dudekwrote: > Hello everyone, > > I would love to help develop Apache Spark. I have run into a (very basic?) > issue which holds me in that mission. > > I followed the `how to contribute` guide, however running ./build/mvn > -DskipTests clean package fails with: > > [INFO] Using zinc server for incremental compilation > [info] 'compiler-interface' not yet compiled for Scala 2.11.8. Compiling... > error: scala.reflect.internal.MissingRequirementError: object > java.lang.Object in compiler mirror not found. > at > scala.reflect.internal.MissingRequirementError$.signal(MissingRequirementError.scala:17) > at > scala.reflect.internal.MissingRequirementError$.notFound(MissingRequirementError.scala:18) > at > scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:53) > > is it perhaps compability issue? Versions I use are as follows: > > ➜ spark git:(master) ✗ ./build/mvn --version > Using `mvn` from path: > /Users/tdudek/Programming/spark/build/apache-maven-3.3.9/bin/mvn > Apache Maven 3.3.9 (bb52d8502b132ec0a5a3f4c09453c07478323dc5; > 2015-11-10T17:41:47+01:00) > Maven home: /Users/tdudek/Programming/spark/build/apache-maven-3.3.9 > Java version: 1.8.0_152, vendor: Oracle Corporation > Java home: > /Library/Java/JavaVirtualMachines/jdk1.8.0_152.jdk/Contents/Home/jre > Default locale: en_PL, platform encoding: US-ASCII > OS name: "mac os x", version: "10.13.1", arch: "x86_64", family: "mac" > > I just lost few hours mindlessly trying to make it work. I hate to waste > other peoples' time and I'm REALLY ashamed at my question, but I think I am > missing something fundamental. > > Cheers, > Tomasz >
Object in compiler mirror not found - maven build
Hello everyone, I would love to help develop Apache Spark. I have run into a (very basic?) issue which holds me in that mission. I followed the `how to contribute` guide, however running ./build/mvn -DskipTests clean package fails with: [INFO] Using zinc server for incremental compilation [info] 'compiler-interface' not yet compiled for Scala 2.11.8. Compiling... error: scala.reflect.internal.MissingRequirementError: object java.lang.Object in compiler mirror not found. at scala.reflect.internal.MissingRequirementError$.signal(MissingRequirementError.scala:17) at scala.reflect.internal.MissingRequirementError$.notFound(MissingRequirementError.scala:18) at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:53) is it perhaps compability issue? Versions I use are as follows: ➜ spark git:(master) ✗ ./build/mvn --version Using `mvn` from path: /Users/tdudek/Programming/spark/build/apache-maven-3.3.9/bin/mvn Apache Maven 3.3.9 (bb52d8502b132ec0a5a3f4c09453c07478323dc5; 2015-11-10T17:41:47+01:00) Maven home: /Users/tdudek/Programming/spark/build/apache-maven-3.3.9 Java version: 1.8.0_152, vendor: Oracle Corporation Java home: /Library/Java/JavaVirtualMachines/jdk1.8.0_152.jdk/Contents/Home/jre Default locale: en_PL, platform encoding: US-ASCII OS name: "mac os x", version: "10.13.1", arch: "x86_64", family: "mac" I just lost few hours mindlessly trying to make it work. I hate to waste other peoples' time and I'm REALLY ashamed at my question, but I think I am missing something fundamental. Cheers, Tomasz
Re: [VOTE] Spark 2.2.1 (RC2)
Yes it downloads recent releases. The test worked for me on a second try, so I suspect a bad mirror. If this comes up frequently we can just add retry logic, as the closer.lua script will return different mirrors each time. The tests all pass for me on the latest Debian, so +1 for this release. (I committed the change to set -Xss4m for tests consistently, but this shouldn't block a release.) On Sat, Nov 25, 2017 at 12:47 PM Felix Cheungwrote: > Ah sorry digging through the history it looks like this is changed > relatively recently and should only download previous releases. > > Perhaps we are intermittently hitting a mirror that doesn’t have the > files? > > > > https://github.com/apache/spark/commit/daa838b8886496e64700b55d1301d348f1d5c9ae > > > On Sat, Nov 25, 2017 at 10:36 AM Felix Cheung > wrote: > >> Thanks Sean. >> >> For the second one, it looks like the >> HiveExternalCatalogVersionsSuite is trying to download the release tgz >> from the official Apache mirror, which won’t work unless the release is >> actually, released? >> >> val preferredMirror = >> Seq("wget", "https://www.apache.org/dyn/closer.lua?preferred=true;, "-q", >> "-O", "-").!!.trim >> val url = s" >> $preferredMirror/spark/spark-$version/spark-$version-bin-hadoop2.7.tgz" >> >> It’s proabbly getting an error page instead. >> >> >> On Sat, Nov 25, 2017 at 10:28 AM Sean Owen wrote: >> >>> I hit the same StackOverflowError as in the previous RC test, but, >>> pretty sure this is just because the increased thread stack size JVM flag >>> isn't applied consistently. This seems to resolve it: >>> >>> https://github.com/apache/spark/pull/19820 >>> >>> This wouldn't block release IMHO. >>> >>> >>> I am currently investigating this failure though -- seems like the >>> mechanism that downloads Spark tarballs needs fixing, or updating, in the >>> 2.2 branch? >>> >>> HiveExternalCatalogVersionsSuite: >>> >>> gzip: stdin: not in gzip format >>> >>> tar: Child returned status 1 >>> >>> tar: Error is not recoverable: exiting now >>> >>> *** RUN ABORTED *** >>> >>> java.io.IOException: Cannot run program "./bin/spark-submit" (in >>> directory "/tmp/test-spark/spark-2.0.2"): error=2, No such file or directory >>> >>> On Sat, Nov 25, 2017 at 12:34 AM Felix Cheung >>> wrote: >>> Please vote on releasing the following candidate as Apache Spark version 2.2.1. The vote is open until Friday December 1, 2017 at 8:00:00 am UTC and passes if a majority of at least 3 PMC +1 votes are cast. [ ] +1 Release this package as Apache Spark 2.2.1 [ ] -1 Do not release this package because ... To learn more about Apache Spark, please see https://spark.apache.org/ The tag to be voted on is v2.2.1-rc2 https://github.com/apache/spark/tree/v2.2.1-rc2 ( e30e2698a2193f0bbdcd4edb884710819ab6397c) List of JIRA tickets resolved in this release can be found here https://issues.apache.org/jira/projects/SPARK/versions/12340470 The release files, including signatures, digests, etc. can be found at: https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-bin/ Release artifacts are signed with the following key: https://dist.apache.org/repos/dist/dev/spark/KEYS The staging repository for this release can be found at: https://repository.apache.org/content/repositories/orgapachespark-1257/ The documentation corresponding to this release can be found at: https://dist.apache.org/repos/dist/dev/spark/spark-2.2.1-rc2-docs/_site/index.html *FAQ* *How can I help test this release?* If you are a Spark user, you can help us test this release by taking an existing Spark workload and running on this release candidate, then reporting any regressions. If you're working in PySpark you can set up a virtual env and install the current RC and see if anything important breaks, in the Java/Scala you can add the staging repository to your projects resolvers and test with the RC (make sure to clean up the artifact cache before/after so you don't end up building with a out of date RC going forward). *What should happen to JIRA tickets still targeting 2.2.1?* Committers should look at those and triage. Extremely important bug fixes, documentation, and API tweaks that impact compatibility should be worked on immediately. Everything else please retarget to 2.2.2. *But my bug isn't fixed!??!* In order to make timely releases, we will typically not hold the release unless the bug in question is a regression from 2.2.0. That being said if there is something which is a regression form 2.2.0 that has not been correctly targeted please ping a committer to help target the issue (you can see the open