For the build I am using java version "1.7.0_65" which seems to be the same
as the one on the spark host. However one is Oracle and the other is
OpenJDK. Does that make any difference?

On Wed, Apr 8, 2015 at 1:24 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> What version of Java do you use to build ?
>
> Cheers
>
> On Wed, Apr 8, 2015 at 12:43 PM, Mohit Anchlia <mohitanch...@gmail.com>
> wrote:
>
>> I am seeing the following, is this because of my maven version?
>>
>> 15/04/08 15:42:22 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0,
>> ip-10-241-251-232.us-west-2.compute.internal):
>> java.io.InvalidClassException: org.apache.spark.Aggregator; local class
>> incompatible: stream classdesc serialVersionUID = 5032037208639381169,
>> local class serialVersionUID = -9085606473104903453
>>         at
>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>>         at
>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>>         at
>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>         at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>         at
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>         at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>         at
>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>         at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>         at
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>
>>
>> <dependency>
>>
>> <groupId>org.apache.spark</groupId>
>>
>> <artifactId>spark-streaming_2.10</artifactId>
>>
>> <version>1.2.0</version>
>>
>> </dependency>
>>
>> <dependency>
>>
>> <groupId>org.apache.spark</groupId>
>>
>> <artifactId>spark-core_2.10</artifactId>
>>
>> <version>1.2.0</version>
>>
>> </dependency>
>>
>>
>

Reply via email to