Dear Gerard,

All servers share the spark binaries through NFS; It is unlikly that other 
servers contains the old class. I will test later with one server and see if I 
got the same problem..

Regards,
Zuhair Khayyat

On Nov 27, 2013, at 6:29 PM, Gerard Maas <gerard.m...@gmail.com> wrote:

> From the looks of your exception, you modified your local class, but you
> forgot to deploy those local changes to the cluster. This error msg:
> classdesc serialVersionUID = 5151096093324583655, local class
> serialVersionUID = 9012954318378784201
> 
> indicates that a version being de-serialized is different from the local
> version. Make sure you deploy your changes across your Spark cluster.
> 
> -kr, Gerard.
> 
> 
> On Wed, Nov 27, 2013 at 4:22 PM, Zuhair Khayyat 
> <zuhair.khay...@gmail.com>wrote:
> 
>> Dear SPARK members,
>> 
>> I am trying to start developing on SPARK source code. I have added a new
>> dummy function in RDD.scala to test if it compiles and runs. The modified
>> Spark compiled correctly but when I execute my code I got the following
>> error:
>> 
>> java.io.InvalidClassException: spark.RDD; local class incompatible: stream
>> classdesc serialVersionUID = 5151096093324583655, local class
>> serialVersionUID = 9012954318378784201
>>        at
>> java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617)
>>        at
>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1620)
>>        at
>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1515)
>>        at
>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1620)
>>        at
>> java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1515)
>>        at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1769)
>>        at
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1348)
>>        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>        at
>> spark.JavaDeserializationStream.readObject(JavaSerializer.scala:23)
>>        at
>> spark.scheduler.ShuffleMapTask$.deserializeInfo(ShuffleMapTask.scala:54)
>>        at
>> spark.scheduler.ShuffleMapTask.readExternal(ShuffleMapTask.scala:111)
>>        at
>> java.io.ObjectInputStream.readExternalData(ObjectInputStream.java:1835)
>>        at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1794)
>>        at
>> java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1348)
>>        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
>>        at
>> spark.JavaDeserializationStream.readObject(JavaSerializer.scala:23)
>>        at
>> spark.JavaSerializerInstance.deserialize(JavaSerializer.scala:45)
>>        at spark.executor.Executor$TaskRunner.run(Executor.scala:96)
>>        at
>> 
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>        at
>> 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>        at java.lang.Thread.run(Thread.java:724)
>> 13/11/27 17:47:43 ERROR executor.StandaloneExecutorBackend: Driver or
>> worker disconnected! Shutting down.
>> 
>> Can you please help me to find out what went wrong? Thank you
>> 
>> Zuhair Khayyat
>> 

Reply via email to