This should be okay, but make sure that your cluster also has the right code
deployed. Maybe you have the wrong one.
If you built Spark from source multiple times, you may also want to try sbt
clean before sbt assembly.
Matei
On August 1, 2014 at 12:00:07 PM, SK (skrishna...@gmail.com) wrote:
Hi,
I upgraded to 1.0.1 from 1.0 a couple of weeks ago and have been able to use
some of the features advertised in 1.0.1. However, I get some compilation
errors in some cases and based on user response, these errors have been
addressed in the 1.0.1 version and so I should not be getting these errors.
So I want to make sure I followed the correct upgrade process as below (I am
running Spark on single machine in standalone mode - so no cluster
deployment):
- set SPARK_HOME to the new version
- run sbt assembly in SPARK_HOME to build the new Spark jars
- in the project sbt file point the libraryDependencies for spark-core and
other libraries to the 1.0.1 version and run sbt assembly to build the
project jar.
Is there anything else I need to do to ensure that no old jars are being
used? For example do I need to manually delete any old jars?
thanks
--
View this message in context:
http://apache-spark-user-list.1001560.n3.nabble.com/correct-upgrade-process-tp11194.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.