I just flailed on this a bit before finding this email. Can someone please update https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools#UsefulDeveloperTools-IDESetup
On Mon, Apr 4, 2016 at 10:01 PM, Reynold Xin <r...@databricks.com> wrote: > pyspark and R > > On Mon, Apr 4, 2016 at 9:59 PM, Marcelo Vanzin <van...@cloudera.com> > wrote: > >> No, tests (except pyspark) should work without having to package anything >> first. >> >> On Mon, Apr 4, 2016 at 9:58 PM, Koert Kuipers <ko...@tresata.com> wrote: >> > do i need to run sbt package before doing tests? >> > >> > On Mon, Apr 4, 2016 at 11:00 PM, Marcelo Vanzin <van...@cloudera.com> >> wrote: >> >> >> >> Hey all, >> >> >> >> We merged SPARK-13579 today, and if you're like me and have your >> >> hands automatically type "sbt assembly" anytime you're building Spark, >> >> that won't work anymore. >> >> >> >> You should now use "sbt package"; you'll still need "sbt assembly" if >> >> you require one of the remaining assemblies (streaming connectors, >> >> yarn shuffle service). >> >> >> >> >> >> -- >> >> Marcelo >> >> >> >> --------------------------------------------------------------------- >> >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> >> For additional commands, e-mail: dev-h...@spark.apache.org >> >> >> > >> >> >> >> -- >> Marcelo >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> For additional commands, e-mail: dev-h...@spark.apache.org >> >> > -- Michael Gummelt Software Engineer Mesosphere