[ 
https://issues.apache.org/jira/browse/SPARK-5351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14285820#comment-14285820
 ] 

Apache Spark commented on SPARK-5351:
-------------------------------------

User 'maropu' has created a pull request for this issue:
https://github.com/apache/spark/pull/4136

> Can't zip RDDs with unequal numbers of partitions in 
> ReplicatedVertexView.upgrade()
> -----------------------------------------------------------------------------------
>
>                 Key: SPARK-5351
>                 URL: https://issues.apache.org/jira/browse/SPARK-5351
>             Project: Spark
>          Issue Type: Bug
>          Components: GraphX
>            Reporter: Takeshi Yamamuro
>
> If the value of 'spark.default.parallelism' does not match the number of 
> partitoins in EdgePartition(EdgeRDDImpl), 
> the following error occurs in ReplicatedVertexView.scala:72;
> object GraphTest extends Logging {
>   def run[VD: ClassTag, ED: ClassTag](graph: Graph[VD, ED]): VertexRDD[Int] = 
> {
>     graph.aggregateMessages[Int](
>       ctx => {
>         ctx.sendToSrc(1)
>         ctx.sendToDst(2)
>       },
>       _ + _)
>   }
> }
> val g = GraphLoader.edgeListFile(sc, "graph.txt")
> val rdd = GraphTest.run(g)
> java.lang.IllegalArgumentException: Can't zip RDDs with unequal numbers of 
> partitions
>       at 
> org.apache.spark.rdd.ZippedPartitionsBaseRDD.getPartitions(ZippedPartitionsRDD.scala:57)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:206)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>       at scala.Option.getOrElse(Option.scala:120)
>       at org.apache.spark.rdd.RDD.partitions(RDD.scala:204)
>       at 
> org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:32)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:206)
>       at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204)
>       at scala.Option.getOrElse(Option.scala:120)
>       at org.apache.spark.rdd.RDD.partitions(RDD.scala:204)
>       at org.apache.spark.ShuffleDependency.<init>(Dependency.scala:82)
>       at 
> org.apache.spark.rdd.ShuffledRDD.getDependencies(ShuffledRDD.scala:80)
>       at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:193)
>       at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:191)
>     ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to