I guess he used client model and the local Spark version is 1.5.2 but the standalone Spark version is 1.5.1. In other words, he used a 1.5.2 driver to talk with 1.5.1 executors.
On Mon, Feb 1, 2016 at 2:08 PM, Holden Karau <hol...@pigscanfly.ca> wrote: > So I'm a little confused to exactly how this might have happened - but one > quick guess is that maybe you've built an assembly jar with Spark core, can > you mark it is a provided and or post your build file? > > On Fri, Jan 29, 2016 at 7:35 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> I logged SPARK-13084 >> >> For the moment, please consider running with 1.5.2 on all the nodes. >> >> On Fri, Jan 29, 2016 at 5:29 AM, Jason Plurad <plur...@gmail.com> wrote: >> >>> I agree with you, Ted, if RDD had a serial version UID this might not be >>> an issue. So that could be a JIRA to submit to help avoid version >>> mismatches in future Spark versions, but that doesn't help my current >>> situation between 1.5.1 and 1.5.2. >>> >>> Any other ideas? Thanks. >>> On Thu, Jan 28, 2016 at 5:06 PM Ted Yu <yuzhih...@gmail.com> wrote: >>> >>>> I am not Scala expert. >>>> >>>> RDD extends Serializable but doesn't have @SerialVersionUID() >>>> annotation. >>>> This may explain what you described. >>>> >>>> One approach is to add @SerialVersionUID so that RDD's have stable >>>> serial version UID. >>>> >>>> Cheers >>>> >>>> On Thu, Jan 28, 2016 at 1:38 PM, Jason Plurad <plur...@gmail.com> >>>> wrote: >>>> >>>>> I've searched through the mailing list archive. It seems that if you >>>>> try to run, for example, a Spark 1.5.2 program against a Spark 1.5.1 >>>>> standalone server, you will run into an exception like this: >>>>> >>>>> WARN org.apache.spark.scheduler.TaskSetManager - Lost task 0.0 in >>>>> stage 0.0 (TID 0, 192.168.14.103): java.io.InvalidClassException: >>>>> org.apache.spark.rdd.RDD; local class incompatible: stream classdesc >>>>> serialVersionUID = -3343649307726848892, local class serialVersionUID = >>>>> -3996494161745401652 >>>>> >>>>> If my application is using a library that builds against Spark 1.5.2, >>>>> does that mean that my application is now tied to that same Spark >>>>> standalone server version? >>>>> >>>>> Is there a recommended way for that library to have a Spark dependency >>>>> but keep it compatible against a wider set of versions, i.e. any version >>>>> 1.5.x? >>>>> >>>>> Thanks! >>>>> >>>> >>>> >> > > > -- > Cell : 425-233-8271 > Twitter: https://twitter.com/holdenkarau >