[ 
https://issues.apache.org/jira/browse/PIG-4693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14947062#comment-14947062
 ] 

Srikanth Sundarrajan commented on PIG-4693:
-------------------------------------------

Yes. After the AM is launched, client makes a request to the AM for 
newHadoopRDD corresponding to the Load Statement, which is shipped through the 
kryo serializer as I had set this as the default spark serializer. The stack 
trace in the original comment has more detail. Posting a section of it here for 
quick reference.
{noformat}
java.lang.NoSuchMethodError: 
com.esotericsoftware.kryo.Kryo.setInstantiatorStrategy(Lorg/objenesis/strategy/InstantiatorStrategy;)V
        at com.twitter.chill.KryoBase.setInstantiatorStrategy(KryoBase.scala:86)
        at 
com.twitter.chill.EmptyScalaKryoInstantiator.newKryo(ScalaKryoInstantiator.scala:59)
        at 
org.apache.spark.serializer.KryoSerializer.newKryo(KryoSerializer.scala:80)
        at 
org.apache.spark.serializer.KryoSerializerInstance.borrowKryo(KryoSerializer.scala:227)
        at 
org.apache.spark.serializer.KryoSerializerInstance.<init>(KryoSerializer.scala:212)
        at 
org.apache.spark.serializer.KryoSerializer.newInstance(KryoSerializer.scala:128)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:201)
        at 
org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:102)
        at 
org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:85)
        at 
org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
        at 
org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:62)
        at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1291)
        at org.apache.spark.rdd.NewHadoopRDD.<init>(NewHadoopRDD.scala:77)
        at 
org.apache.spark.SparkContext$$anonfun$newAPIHadoopRDD$1.apply(SparkContext.scala:1099)
        at 
org.apache.spark.SparkContext$$anonfun$newAPIHadoopRDD$1.apply(SparkContext.scala:1094)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108)
        at org.apache.spark.SparkContext.withScope(SparkContext.scala:681)
        at 
org.apache.spark.SparkContext.newAPIHadoopRDD(SparkContext.scala:1094)
        at 
org.apache.pig.backend.hadoop.executionengine.spark.converter.LoadConverter.convert(LoadConverter.java:88)
{noformat}

> Class conflicts: Kryo bundled in spark vs kryo bundled with pig
> ---------------------------------------------------------------
>
>                 Key: PIG-4693
>                 URL: https://issues.apache.org/jira/browse/PIG-4693
>             Project: Pig
>          Issue Type: Sub-task
>          Components: spark
>    Affects Versions: spark-branch
>            Reporter: Srikanth Sundarrajan
>            Assignee: Srikanth Sundarrajan
>              Labels: spork
>             Fix For: spark-branch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to