After I do a clean rebuild. It works now. Thanks, Qiuzhuang
On Sat, Oct 25, 2014 at 9:42 AM, Nan Zhu <zhunanmcg...@gmail.com> wrote: > According to my experience, there are more issues rather than > BlockManager when you try to run spark application whose build version is > different with your cluster…. > > I once tried to make jdbc server build with branch-jdbc-1.0 run with a > branch-1.0 cluster…no workaround exits…just had to replace cluster jar with > branch-jdbc-1.0 jar file….. > > Best, > > -- > Nan Zhu > > On Friday, October 24, 2014 at 9:23 PM, Josh Rosen wrote: > > Are all processes (Master, Worker, Executors, Driver) running the same > Spark build? This error implies that you’re seeing protocol / binary > incompatibilities between your Spark driver and cluster. > > Spark is API-compatibile across the 1.x series, but we don’t make binary > link-level compatibility guarantees: > https://cwiki.apache.org/confluence/display/SPARK/Spark+Versioning+Policy. > This means that your Spark driver’s runtime classpath should use the same > version of Spark that’s installed on your cluster. You can compile against > a different API-compatible version of Spark, but the runtime versions must > match across all components. > > To fix this issue, I’d check that you’ve run the “package” and “assembly” > phases and that your Spark cluster is using this updated version. > > - Josh > > On October 24, 2014 at 6:17:26 PM, Qiuzhuang Lian ( > qiuzhuang.l...@gmail.com) wrote: > > Hi, > > I update git today and when connecting to spark cluster, I got > the serialVersionUID incompatible error in class BlockManagerId. > > Here is the log, > > Shouldn't we better give BlockManagerId a constant serialVersionUID avoid > this? > > Thanks, > Qiuzhuang > > scala> val rdd = sc.parparallelize(1 to 100014/10/25 09:10:48 ERROR > Remoting: org.apache.spark.storage.BlockManagerId; local class > incompatible: stream classdesc serialVersionUID = 2439208141545036836, > local class serialVersionUID = 4657685702603429489 > java.io.InvalidClassException: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > akka.serialization.JavaSerializer$$anonfun$1.apply(Serializer.scala:136) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) > at > akka.serialization.JavaSerializer.fromBinary(Serializer.scala:136) > at > akka.serialization.Serialization$$anonfun$deserialize$1.apply(Serialization.scala:104) > > at scala.util.Try$.apply(Try.scala:161) > at > akka.serialization.Serialization.deserialize(Serialization.scala:98) > at > akka.remote.MessageSerializer$.deserialize(MessageSerializer.scala:23) > at > akka.remote.DefaultMessageDispatcher.payload$lzycompute$1(Endpoint.scala:58) > > at akka.remote.DefaultMessageDispatcher.payload$1(Endpoint.scala:58) > at akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:76) > at > akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:937) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:415) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393) > > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 14/10/25 09:10:48 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 1 > 0014/10/25 09:11:21 ERROR Remoting: > org.apache.spark.storage.BlockManagerId; local class incompatible: stream > classdesc serialVersionUID = 2439208141545036836, local class > serialVersionUID = 4657685702603429489 > java.io.InvalidClassException: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > akka.serialization.JavaSerializer$$anonfun$1.apply(Serializer.scala:136) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) > at > akka.serialization.JavaSerializer.fromBinary(Serializer.scala:136) > at > akka.serialization.Serialization$$anonfun$deserialize$1.apply(Serialization.scala:104) > > at scala.util.Try$.apply(Try.scala:161) > at > akka.serialization.Serialization.deserialize(Serialization.scala:98) > at > akka.remote.MessageSerializer$.deserialize(MessageSerializer.scala:23) > at > akka.remote.DefaultMessageDispatcher.payload$lzycompute$1(Endpoint.scala:58) > > at akka.remote.DefaultMessageDispatcher.payload$1(Endpoint.scala:58) > at akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:76) > at > akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:937) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:415) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393) > > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 14/10/25 09:11:21 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 1 > 14/10/25 09:11:54 INFO SparkDeploySchedulerBackend: Registered executor: > Actor[akka.tcp://sparkexecu...@dev-02.springb.gz:50006/user/Executor#-1410691203] > > with ID 1 > 14/10/25 09:11:54 INFO DAGScheduler: Host added was in lost list earlier: > DEV-02.SpringB.GZ > 14/10/25 09:11:55 ERROR TaskSchedulerImpl: Lost executor 1 on > DEV-02.SpringB.GZ: remote Akka client disassociated > 14/10/25 09:11:55 WARN ReliableDeliverySupervisor: Association with remote > system [akka.tcp://sparkexecu...@dev-02.springb.gz:50006] has failed, > address is now gated for [5000] ms. Reason is: [Association failed with > [akka.tcp://sparkexecu...@dev-02.springb.gz:50006]]. > 14/10/25 09:11:55 INFO DAGScheduler: Executor lost: 1 (epoch 1) > 14/10/25 09:11:55 INFO BlockManagerMasterActor: Trying to remove executor > 1 > from BlockManagerMaster. > 14/10/25 09:11:55 INFO BlockManagerMaster: Removed 1 successfully in > removeExecutor > 14/10/25 09:11:55 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 1 > 14/10/25 09:11:55 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 1 > 14/10/25 09:11:55 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/1 is now EXITED (Command exited with code 1) > 14/10/25 09:11:55 INFO SparkDeploySchedulerBackend: Executor > app-20141025091012-0002/1 removed: Command exited with code 1 > 14/10/25 09:11:55 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 1 > 14/10/25 09:11:55 INFO AppClient$ClientActor: Executor added: > app-20141025091012-0002/3 on worker-20141025170311-DEV-02.SpringB.GZ-35162 > (DEV-02.SpringB.GZ:35162) with 2 cores > 14/10/25 09:11:55 INFO SparkDeploySchedulerBackend: Granted executor ID > app-20141025091012-0002/3 on hostPort DEV-02.SpringB.GZ:35162 with 2 > cores, > 512.0 MB RAM > 14/10/25 09:11:55 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/3 is now LOADING > 14/10/25 09:11:55 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/3 is now RUNNING > 14/10/25 09:11:58 INFO SparkDeploySchedulerBackend: Registered executor: > Actor[akka.tcp://sparkexecu...@dev-02.springb.gz:50740/user/Executor#1229699385] > > with ID 3 > 14/10/25 09:11:58 WARN ReliableDeliverySupervisor: Association with remote > system [akka.tcp://sparkexecu...@dev-02.springb.gz:50740] has failed, > address is now gated for [5000] ms. Reason is: > [org.apache.spark.storage.BlockManagerId; local class incompatible: stream > classdesc serialVersionUID = 2439208141545036836, local class > serialVersionUID = 4657685702603429489]. > 14/10/25 09:11:58 ERROR TaskSchedulerImpl: Lost executor 3 on > DEV-02.SpringB.GZ: remote Akka client disassociated > 14/10/25 09:11:58 INFO DAGScheduler: Executor lost: 3 (epoch 2) > 14/10/25 09:11:58 INFO BlockManagerMasterActor: Trying to remove executor > 3 > from BlockManagerMaster. > 14/10/25 09:11:58 INFO BlockManagerMaster: Removed 3 successfully in > removeExecutor > 14/10/25 09:12:31 ERROR Remoting: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > java.io.InvalidClassException: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > akka.serialization.JavaSerializer$$anonfun$1.apply(Serializer.scala:136) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) > at > akka.serialization.JavaSerializer.fromBinary(Serializer.scala:136) > at > akka.serialization.Serialization$$anonfun$deserialize$1.apply(Serialization.scala:104) > > at scala.util.Try$.apply(Try.scala:161) > at > akka.serialization.Serialization.deserialize(Serialization.scala:98) > at > akka.remote.MessageSerializer$.deserialize(MessageSerializer.scala:23) > at > akka.remote.DefaultMessageDispatcher.payload$lzycompute$1(Endpoint.scala:58) > > at akka.remote.DefaultMessageDispatcher.payload$1(Endpoint.scala:58) > at akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:76) > at > akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:937) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:415) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393) > > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 14/10/25 09:12:31 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 3 > 14/10/25 09:13:04 ERROR Remoting: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > java.io.InvalidClassException: org.apache.spark.storage.BlockManagerId; > local class incompatible: stream classdesc serialVersionUID = > 2439208141545036836, local class serialVersionUID = 4657685702603429489 > at > java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:617) > at > java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1622) > at > java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990) > at > java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798) > at > java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370) > at > akka.serialization.JavaSerializer$$anonfun$1.apply(Serializer.scala:136) > at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57) > at > akka.serialization.JavaSerializer.fromBinary(Serializer.scala:136) > at > akka.serialization.Serialization$$anonfun$deserialize$1.apply(Serialization.scala:104) > > at scala.util.Try$.apply(Try.scala:161) > at > akka.serialization.Serialization.deserialize(Serialization.scala:98) > at > akka.remote.MessageSerializer$.deserialize(MessageSerializer.scala:23) > at > akka.remote.DefaultMessageDispatcher.payload$lzycompute$1(Endpoint.scala:58) > > at akka.remote.DefaultMessageDispatcher.payload$1(Endpoint.scala:58) > at akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:76) > at > akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:937) > > at akka.actor.Actor$class.aroundReceive(Actor.scala:465) > at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:415) > at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) > at akka.actor.ActorCell.invoke(ActorCell.scala:487) > at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238) > at akka.dispatch.Mailbox.run(Mailbox.scala:220) > at > akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393) > > at > scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) > at > scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) > > at > scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) > at > scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) > > 14/10/25 09:13:04 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 3 > 14/10/25 09:13:37 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 3 > 14/10/25 09:13:37 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 3 > 14/10/25 09:13:37 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/3 is now EXITED (Command exited with code 1) > 14/10/25 09:13:37 INFO SparkDeploySchedulerBackend: Executor > app-20141025091012-0002/3 removed: Command exited with code 1 > 14/10/25 09:13:37 ERROR SparkDeploySchedulerBackend: Asked to remove non > existant executor 3 > 14/10/25 09:13:37 INFO AppClient$ClientActor: Executor added: > app-20141025091012-0002/4 on worker-20141025170311-DEV-02.SpringB.GZ-35162 > (DEV-02.SpringB.GZ:35162) with 2 cores > 14/10/25 09:13:37 INFO SparkDeploySchedulerBackend: Granted executor ID > app-20141025091012-0002/4 on hostPort DEV-02.SpringB.GZ:35162 with 2 > cores, > 512.0 MB RAM > 14/10/25 09:13:37 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/4 is now LOADING > 14/10/25 09:13:38 INFO AppClient$ClientActor: Executor updated: > app-20141025091012-0002/4 is now RUNNING > 14/10/25 09:13:40 INFO SparkDeploySchedulerBackend: Registered executor: > Actor[akka.tcp://sparkexecu...@dev-02.springb.gz:56019/user/Executor#1354626597] > > with ID 4 > 14/10/25 09:13:40 WARN ReliableDeliverySupervisor: Association with remote > system [akka.tcp://sparkexecu...@dev-02.springb.gz:56019] has failed, > address is now gated for [5000] ms. Reason is: > [org.apache.spark.storage.BlockManagerId; local class incompatible: stream > classdesc serialVersionUID = 2439208141545036836, local class > serialVersionUID = 4657685702603429489]. > 14/10/25 09:13:40 ERROR TaskSchedulerImpl: Lost executor 4 on > DEV-02.SpringB.GZ: remote Akka client disassociated > 14/10/25 09:13:40 INFO DAGScheduler: Executor lost: 4 (epoch 3) > 14/10/25 09:13:40 INFO BlockManagerMasterActor: Trying to remove executor > 4 > from BlockManagerMaster. > 14/10/25 09:13:40 INFO BlockManagerMaster: Removed 4 successfully in > removeExecutor > > >