[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-05-20 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553258#comment-14553258
 ] 

ASF GitHub Bot commented on SPARK-6069:
---

Github user pferrel closed the pull request at:

https://github.com/apache/mahout/pull/125


 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-05-02 Thread Pat Ferrel (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14525267#comment-14525267
 ] 

Pat Ferrel commented on SPARK-6069:
---

Same with Mahout where we were using Guava. Using Scala collections only will 
solve this for us.

The way to work around this is to use the spark.executor.extraClassPath, which 
should point to the correct jar in the native filesystem on every worker! So 
you have to move a dependency jar to every worker. 


 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-05-01 Thread Russell Alexander Spitzer (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14524372#comment-14524372
 ] 

Russell Alexander Spitzer commented on SPARK-6069:
--

Running with --conf spark.files.userClassPathFirst=true yields a different error

{code}
scala cc.sql(SELECT * FROM test.fun as a JOIN test.fun as b ON (a.k = 
b.v)).collect
15/05/01 17:24:34 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0, 
10.0.2.15): java.lang.NoClassDefFoundError: org/apache/spark/Partition
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at 
org.apache.spark.executor.ChildExecutorURLClassLoader$userClassLoader$.findClass(ExecutorURLClassLoader.scala:42)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClass(ClassLoader.java:800)
at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:449)
at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at 
org.apache.spark.executor.ChildExecutorURLClassLoader$userClassLoader$.findClass(ExecutorURLClassLoader.scala:42)
at 
org.apache.spark.executor.ChildExecutorURLClassLoader.findClass(ExecutorURLClassLoader.scala:50)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:412)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at 
org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.scala:30)
at 
org.apache.spark.repl.ExecutorClassLoader$$anonfun$findClass$1.apply(ExecutorClassLoader.scala:57)
at 
org.apache.spark.repl.ExecutorClassLoader$$anonfun$findClass$1.apply(ExecutorClassLoader.scala:57)
at scala.Option.getOrElse(Option.scala:120)
at 
org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:57)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:274)
at 
org.apache.spark.serializer.JavaDeserializationStream$$anon$1.resolveClass(JavaSerializer.scala:59)
at 
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612)
at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
at 
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
at 
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:182)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.ClassNotFoundException: org.apache.spark.Partition
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at 

[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-05-01 Thread Russell Alexander Spitzer (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14524363#comment-14524363
 ] 

Russell Alexander Spitzer commented on SPARK-6069:
--

We've seen the same issue while developing the Spark Cassandra Connector. 
Unless the connector lib is loaded via spark.executor.extraClassPath, 
kryoSerializaition for joins always returns a classNotFound even though all 
operations which don't require a shuffle are fine. 

{code}
com.esotericsoftware.kryo.KryoException: Unable to find class: 
org.apache.spark.sql.cassandra.CassandraSQLRow
at 
com.esotericsoftware.kryo.util.DefaultClassResolver.readName(DefaultClassResolver.java:138)
at 
com.esotericsoftware.kryo.util.DefaultClassResolver.readClass(DefaultClassResolver.java:115)
at com.esotericsoftware.kryo.Kryo.readClass(Kryo.java:610)
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:721)
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:42)
at com.twitter.chill.Tuple2Serializer.read(TupleSerializers.scala:33)
at com.esotericsoftware.kryo.Kryo.readClassAndObject(Kryo.java:732)
at 
org.apache.spark.serializer.KryoDeserializationStream.readObject(KryoSerializer.scala:144)
at 
org.apache.spark.serializer.DeserializationStream$$anon$1.getNext(Serializer.scala:133)
at org.apache.spark.util.NextIterator.hasNext(NextIterator.scala:71)
at 
org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
at 
org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:32)
at 
org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:39)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327)
at 
org.apache.spark.sql.execution.joins.HashedRelation$.apply(HashedRelation.scala:80)
at 
org.apache.spark.sql.execution.joins.ShuffledHashJoin$$anonfun$execute$1.apply(ShuffledHashJoin.scala:46)
at 
org.apache.spark.sql.execution.joins.ShuffledHashJoin$$anonfun$execute$1.apply(ShuffledHashJoin.scala:45)
at 
org.apache.spark.rdd.ZippedPartitionsRDD2.compute(ZippedPartitionsRDD.scala:88)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:280)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:247)
at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:280)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:247)
at org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:280)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:247)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
at org.apache.spark.scheduler.Task.run(Task.scala:56)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:200)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
{code}

Adding the jar to executorExtraClasspath rather than --jars solves the issue.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-04-30 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14521542#comment-14521542
 ] 

ASF GitHub Bot commented on SPARK-6069:
---

Github user pferrel commented on the pull request:

https://github.com/apache/mahout/pull/125#issuecomment-97811136
  
Still would like an eye on BiMap and BiDictionary.

The remaining question here is if it solves 
https://issues.apache.org/jira/browse/SPARK-6069 on a cluster.

I'll concentrate on finding a way to test this. A user tried this on a 
cluster and got an infinite recursion so wouldn't want to push yet.



 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-04-30 Thread Pat Ferrel (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14521664#comment-14521664
 ] 

Pat Ferrel commented on SPARK-6069:
---

didn't mean for those comments to be cross-posted.

Removing the use of Javaserializer to work around this problem in Spark 1.2, 
which is in wide use.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-04-29 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14519940#comment-14519940
 ] 

ASF GitHub Bot commented on SPARK-6069:
---

GitHub user pferrel opened a pull request:

https://github.com/apache/mahout/pull/125

Spark 1.2

Created a Scala based BiMap and BiDictionary and removed usage of Guava 
HashBiMap (actually no Guava is used now), I hope this gets past the 
Javaserializer bug in https://issues.apache.org/jira/browse/SPARK-6069

Also found an error that did an unneeded collect, which will drastically 
reduce memory requirements in reading in data for cooccurrence.

Passing around ```Option[BiDictionary] = None``` instead of an empty 
dictionary, which seems a better idiom when using an immutable BiDictionary.

Questions:
* Passes local build tests but haven't tried on a cluster to verify it 
works in spite of SPARK-6069
* Need to try it on epinions data before pushing
* Where should this go, it's in 0.11.0-SNAPSHOT now
* Still need to try putting Scala libs in mahout shell module only. This is 
to check to see if they are needed in non-shell Scala modules

I'm having trouble getting HDFS+Spark running, even standalone, so if 
anyone wants to try this I can provide instructions.



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/pferrel/mahout spark-1.2

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/mahout/pull/125.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #125


commit 71165a54b9bc99f23c2cce6e8c63720b7fc64001
Author: Andrew Palumbo apalu...@apache.org
Date:   2015-02-24T22:54:15Z

added scala dependencies in math-scala to fix spark-shell

commit 75054c195ed71fe51f6e3de85e598f6ce6d51c3a
Author: pferrel p...@occamsmachete.com
Date:   2015-04-29T14:36:58Z

initial cut at making cooccurrence work on Spark 1.2--runs locally but 
cleanup needed, also major memory use optimizations

commit 990347b16258b3bf410c130314a2225654100181
Author: pferrel p...@occamsmachete.com
Date:   2015-04-29T15:49:24Z

refactored BiDictionary and BiMap to carve off dicictionary specific 
behavior in the companion object

commit 1d897176e4d2e50552d476741e5c314c072f
Author: pferrel p...@occamsmachete.com
Date:   2015-04-29T16:19:08Z

runs build tests now removing the .create factory method, which will 
require creation only when keys are already available

commit f3611855fbf2b06bc3faa7083484d426e84e1b73
Author: pferrel p...@occamsmachete.com
Date:   2015-04-29T18:37:38Z

using Option[BiDictionary] = None now instread of passing around an empty 
BiDictionary




 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-03-02 Thread Pat Ferrel (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14343372#comment-14343372
 ] 

Pat Ferrel commented on SPARK-6069:
---

ok, I'll try 1.3 as soon as I get a chance.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-03-02 Thread Marcelo Vanzin (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14343437#comment-14343437
 ] 

Marcelo Vanzin commented on SPARK-6069:
---

I think this is a dupe of SPARK-5470, but haven't really tested. SPARK-4660 is 
a similar (same?) issue for the Java serializer.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-03-02 Thread Pat Ferrel (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14343597#comment-14343597
 ] 

Pat Ferrel commented on SPARK-6069:
---

great since it's fixed in 1.3 I'll definitely try that next and resolve this if 
it flies.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-02-28 Thread Sean Owen (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14341819#comment-14341819
 ] 

Sean Owen commented on SPARK-6069:
--

No I set it kind of preemptively. I don't know that I serialize any Guava 
classes though, come to think of it.
I am using YARN + Hadoop 2.5.

I don't think it should be necessary in general. Guava is a strange special 
case, so though it worth trying.

If you have the energy, you might try 1.3.0-SNAPSHOT since I see a few things 
fixed that may be relevant:
https://issues.apache.org/jira/browse/SPARK-4877
https://issues.apache.org/jira/browse/SPARK-4660


 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-6069) Deserialization Error ClassNotFoundException with Kryo, Guava 14

2015-02-28 Thread Pat Ferrel (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-6069?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14341809#comment-14341809
 ] 

Pat Ferrel commented on SPARK-6069:
---

Embarrassed to say still on Hadoop 1.2.1 and so no yarn. The packaging is not 
in the app jar but a separate pruned down dependencies-only jar. I can see why 
yarn would throw a unique kink into the situation.  So I guess you ran into 
this and had to use the {{user.classpath.first}} work around or are you saying 
it doesn't occur in oryx?

Still none of this should be necessary, right? Why else would jars be specified 
in to context creation? We do have a work around if someone has to work with 
1.2.1 but because of that it doesn't seem like a good version to recommend. 
Maybe I'll try 1.2 and install H2 and yarn--which seems like what the distros 
support.

 Deserialization Error ClassNotFoundException with Kryo, Guava 14
 

 Key: SPARK-6069
 URL: https://issues.apache.org/jira/browse/SPARK-6069
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 1.2.1
 Environment: Standalone one worker cluster on localhost, or any 
 cluster
Reporter: Pat Ferrel
Priority: Critical

 A class is contained in the jars passed in when creating a context. It is 
 registered with kryo. The class (Guava HashBiMap) is created correctly from 
 an RDD and broadcast but the deserialization fails with ClassNotFound.
 The work around is to hard code the path to the jar and make it available on 
 all workers. Hard code because we are creating a library so there is no easy 
 way to pass in to the app something like:
 spark.executor.extraClassPath  /path/to/some.jar



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org