Hi, I don't think there are any sbt-related changes in Spark 2.0. Just different versions in libraryDependencies.
As to the article, I'm surprised it didn't mention using sbt-assembly [1] for docker-like deployment or sbt-native-packager [2] that could create a Docker image. [1] https://github.com/sbt/sbt-assembly [2] http://www.scala-sbt.org/sbt-native-packager/ Pozdrawiam, Jacek Laskowski ---- https://medium.com/@jaceklaskowski/ Mastering Apache Spark http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Tue, Jul 26, 2016 at 2:54 PM, Martin Somers <sono...@gmail.com> wrote: > Just wondering > > Whats is the correct way of building a spark job using scala - are there any > changes coming with spark v2 > > Ive been following this post > > http://www.infoobjects.com/spark-submit-with-sbt/ > > > > Then again Ive been mainly using docker locally what is decent container for > submitting these jobs locally > > Im getting to a stage where I need to submit jobs remotely and thinking of > the best way of doing so > > > tks > > M --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org