Thank you Marcelo. I tried your suggestion (# mvn -pl :spark-examples_2.10 compile), but it required to download many spark components (as listed below), which I have already compiled on my server.
Downloading: https://repo1.maven.org/maven2/org/apache/spark/spark-core_2.10/1.1.0/spark-core_2.10-1.1.0.pom ... Downloading: https://repo1.maven.org/maven2/org/apache/spark/spark-streaming_2.10/1.1.0/spark-streaming_2.10-1.1.0.pom ... Downloading: https://repository.jboss.org/nexus/content/repositories/releases/org/apache/spark/spark-hive_2.10/1.1.0/spark-hive_2.10-1.1.0.pom ... This problem didn't happen when I compiled the whole project using ``mvn -DskipTests package''. I guess some configurations have to be made to tell mvn the dependencies are local. Any idea for that? Thank you for your help! Cheers, Yiming -----邮件原件----- 发件人: Marcelo Vanzin [mailto:van...@cloudera.com] 发送时间: 2014年11月16日 10:26 收件人: sdi...@gmail.com 抄送: user@spark.apache.org 主题: Re: How to incrementally compile spark examples using mvn I haven't tried scala:cc, but you can ask maven to just build a particular sub-project. For example: mvn -pl :spark-examples_2.10 compile On Sat, Nov 15, 2014 at 5:31 PM, Yiming (John) Zhang <sdi...@gmail.com> wrote: > Hi, > > > > I have already successfully compile and run spark examples. My problem > is that if I make some modifications (e.g., on SparkPi.scala or > LogQuery.scala) I have to use “mvn -DskipTests package” to rebuild the > whole spark project and wait a relatively long time. > > > > I also tried “mvn scala:cc” as described in > http://spark.apache.org/docs/latest/building-with-maven.html, but I > could only get infinite stop like: > > [INFO] --- scala-maven-plugin:3.2.0:cc (default-cli) @ spark-parent > --- > > [INFO] wait for files to compile... > > > > Is there any method to incrementally compile the examples using mvn? > Thank you! > > > > Cheers, > > Yiming -- Marcelo --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org