ah yah that would not work. On Mon, Feb 1, 2016 at 2:31 PM, Shixiong(Ryan) Zhu <shixi...@databricks.com> wrote:
> I guess he used client model and the local Spark version is 1.5.2 but the > standalone Spark version is 1.5.1. In other words, he used a 1.5.2 driver > to talk with 1.5.1 executors. > > On Mon, Feb 1, 2016 at 2:08 PM, Holden Karau <hol...@pigscanfly.ca> wrote: > >> So I'm a little confused to exactly how this might have happened - but >> one quick guess is that maybe you've built an assembly jar with Spark core, >> can you mark it is a provided and or post your build file? >> >> On Fri, Jan 29, 2016 at 7:35 AM, Ted Yu <yuzhih...@gmail.com> wrote: >> >>> I logged SPARK-13084 >>> >>> For the moment, please consider running with 1.5.2 on all the nodes. >>> >>> On Fri, Jan 29, 2016 at 5:29 AM, Jason Plurad <plur...@gmail.com> wrote: >>> >>>> I agree with you, Ted, if RDD had a serial version UID this might not >>>> be an issue. So that could be a JIRA to submit to help avoid version >>>> mismatches in future Spark versions, but that doesn't help my current >>>> situation between 1.5.1 and 1.5.2. >>>> >>>> Any other ideas? Thanks. >>>> On Thu, Jan 28, 2016 at 5:06 PM Ted Yu <yuzhih...@gmail.com> wrote: >>>> >>>>> I am not Scala expert. >>>>> >>>>> RDD extends Serializable but doesn't have @SerialVersionUID() >>>>> annotation. >>>>> This may explain what you described. >>>>> >>>>> One approach is to add @SerialVersionUID so that RDD's have stable >>>>> serial version UID. >>>>> >>>>> Cheers >>>>> >>>>> On Thu, Jan 28, 2016 at 1:38 PM, Jason Plurad <plur...@gmail.com> >>>>> wrote: >>>>> >>>>>> I've searched through the mailing list archive. It seems that if you >>>>>> try to run, for example, a Spark 1.5.2 program against a Spark 1.5.1 >>>>>> standalone server, you will run into an exception like this: >>>>>> >>>>>> WARN org.apache.spark.scheduler.TaskSetManager - Lost task 0.0 in >>>>>> stage 0.0 (TID 0, 192.168.14.103): java.io.InvalidClassException: >>>>>> org.apache.spark.rdd.RDD; local class incompatible: stream classdesc >>>>>> serialVersionUID = -3343649307726848892, local class serialVersionUID = >>>>>> -3996494161745401652 >>>>>> >>>>>> If my application is using a library that builds against Spark 1.5.2, >>>>>> does that mean that my application is now tied to that same Spark >>>>>> standalone server version? >>>>>> >>>>>> Is there a recommended way for that library to have a Spark >>>>>> dependency but keep it compatible against a wider set of versions, i.e. >>>>>> any >>>>>> version 1.5.x? >>>>>> >>>>>> Thanks! >>>>>> >>>>> >>>>> >>> >> >> >> -- >> Cell : 425-233-8271 >> Twitter: https://twitter.com/holdenkarau >> > > -- Cell : 425-233-8271 Twitter: https://twitter.com/holdenkarau