bq. one is Oracle and the other is OpenJDK

I don't have experience with mixed JDK's.
Can you try with using single JDK ?

Cheers

On Wed, Apr 8, 2015 at 3:26 PM, Mohit Anchlia <mohitanch...@gmail.com>
wrote:

> For the build I am using java version "1.7.0_65" which seems to be the
> same as the one on the spark host. However one is Oracle and the other is
> OpenJDK. Does that make any difference?
>
> On Wed, Apr 8, 2015 at 1:24 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> What version of Java do you use to build ?
>>
>> Cheers
>>
>> On Wed, Apr 8, 2015 at 12:43 PM, Mohit Anchlia <mohitanch...@gmail.com>
>> wrote:
>>
>>> I am seeing the following, is this because of my maven version?
>>>
>>> 15/04/08 15:42:22 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID
>>> 0, ip-10-241-251-232.us-west-2.compute.internal):
>>> java.io.InvalidClassException: org.apache.spark.Aggregator; local class
>>> incompatible: stream classdesc serialVersionUID = 5032037208639381169,
>>> local class serialVersionUID = -9085606473104903453
>>>         at
>>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>>>         at
>>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622)
>>>         at
>>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
>>>         at
>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
>>>         at
>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>         at
>>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
>>>         at
>>> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
>>>         at
>>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
>>>         at
>>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
>>>
>>>
>>> <dependency>
>>>
>>> <groupId>org.apache.spark</groupId>
>>>
>>> <artifactId>spark-streaming_2.10</artifactId>
>>>
>>> <version>1.2.0</version>
>>>
>>> </dependency>
>>>
>>> <dependency>
>>>
>>> <groupId>org.apache.spark</groupId>
>>>
>>> <artifactId>spark-core_2.10</artifactId>
>>>
>>> <version>1.2.0</version>
>>>
>>> </dependency>
>>>
>>>
>>
>

Reply via email to