The master SBT builds seem OK, like:
https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test%20(Dashboard)/job/spark-master-test-sbt-hadoop-2.7/

It looks like an issue between Windows, SBT, and your env I think.

On Wed, Feb 7, 2018 at 5:12 PM ds <davor.su...@gmail.com> wrote:

> After cloning today's version of spark-master, I run the following command:
> S:\spark-master>sbt ./build/sbt -Pyarn -Phadoop-2.7 -Dhadoop.version=2.7.0
> -Phive -Phive-thriftserver clean package
> with the intention of building both the source and test projects and
> generating the corresponding .jar files.
>
> The script started regularly, but ultimately failed with the following log
> excerpt:
> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
> MaxPermSize=256m;
> support was removed in
> 8.0
> [info] Loading project definition from S:\spark-master\project
> [info] Resolving key references (16939 settings) ...
> [info] Set current project to spark-parent (in build
> file:/S:/spark-master/)
> [error] Expected letter
> [error] Expected symbol
> [error] Expected '!'
> [error] Expected '+'
> [error] Expected '++'
> [error] Expected '^'
> [error] Expected '^^'
> [error] Expected 'debug'
> [error] Expected 'info'
> [error] Expected 'warn'
> [error] Expected 'error'
> [error] Expected ';'
> [error] Expected end of input.
> [error] Expected 'early('
> [error] Expected '-'
> [error] Expected '--'
> [error] Expected 'show'
> [error] Expected 'all'
> [error] Expected '*'
> [error] Expected '{'
> [error] Expected project ID
> [error] Expected configuration
> [error] Expected key
> [error] ./build/sbt
> [error] ^
>
>
> I tried to follow the instructions found at
> http://www.sparktutorials.net/building-apache-spark-on-your-local-machine
> to
> the best of my understanding, but I don't know how to interpret the error
> and where to begin the troubleshooting.
>
> I'm using eclipse as my IDE, so both scala and java seem to be setup
> properly. Although, after running out of options, I simply run
> S:\spark-master>sbt compile, which failed with these errors (I don't know
> whether this is relevant):
> [error] (core/compile:managedResources) java.io.IOException: Cannot run
> program "bash": CreateProcess error=2, The system cannot find the file
> specified
> [error] (network-common/compile:compileIncremental) java.io.IOException:
> Cannot run program "S:\Program Files\Java\bin\javac" (in directory
> "S:\spark-master"): CreateProcess error=2, The system cannot find the file
> specified
> [error] (tags/compile:compileIncremental) java.io.IOException: Cannot run
> program "S:\Program Files\Java\bin\javac" (in directory "S:\spark-master"):
> CreateProcess error=2, The system cannot find the file specified
>
> Note that javac is located in S:\Program Files\Java\jdk1.8.0_77\bin.
>
> So, I would appreciate help in building and packaging the src and test
> components of the spark source.
>
>
>
>
>
>
> --
> Sent from: http://apache-spark-developers-list.1001551.n3.nabble.com/
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to