Thank you Marcelo. I tried your suggestion (# mvn -pl :spark-examples_2.10 
compile), but it required to download many spark components (as listed below), 
which I have already compiled on my server.

Downloading: 
https://repo1.maven.org/maven2/org/apache/spark/spark-core_2.10/1.1.0/spark-core_2.10-1.1.0.pom
...
Downloading: 
https://repo1.maven.org/maven2/org/apache/spark/spark-streaming_2.10/1.1.0/spark-streaming_2.10-1.1.0.pom
...
Downloading: 
https://repository.jboss.org/nexus/content/repositories/releases/org/apache/spark/spark-hive_2.10/1.1.0/spark-hive_2.10-1.1.0.pom
...

This problem didn't happen when I compiled the whole project using ``mvn 
-DskipTests package''. I guess some configurations have to be made to tell mvn 
the dependencies are local. Any idea for that?

Thank you for your help!

Cheers,
Yiming

-----邮件原件-----
发件人: Marcelo Vanzin [mailto:van...@cloudera.com] 
发送时间: 2014年11月16日 10:26
收件人: sdi...@gmail.com
抄送: user@spark.apache.org
主题: Re: How to incrementally compile spark examples using mvn

I haven't tried scala:cc, but you can ask maven to just build a particular 
sub-project. For example:

  mvn -pl :spark-examples_2.10 compile

On Sat, Nov 15, 2014 at 5:31 PM, Yiming (John) Zhang <sdi...@gmail.com> wrote:
> Hi,
>
>
>
> I have already successfully compile and run spark examples. My problem 
> is that if I make some modifications (e.g., on SparkPi.scala or 
> LogQuery.scala) I have to use “mvn -DskipTests package” to rebuild the 
> whole spark project and wait a relatively long time.
>
>
>
> I also tried “mvn scala:cc” as described in 
> http://spark.apache.org/docs/latest/building-with-maven.html, but I 
> could only get infinite stop like:
>
> [INFO] --- scala-maven-plugin:3.2.0:cc (default-cli) @ spark-parent 
> ---
>
> [INFO] wait for files to compile...
>
>
>
> Is there any method to incrementally compile the examples using mvn? 
> Thank you!
>
>
>
> Cheers,
>
> Yiming



--
Marcelo


---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to