[
https://issues.apache.org/jira/browse/SPARK-2610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Yin Huai updated SPARK-2610:
----------------------------
Summary: When spark.serializer is set as
org.apache.spark.serializer.KryoSerializer, importing a method causes multiple
spark applications creations (was: When spark.serializer is set as
org.apache.spark.serializer.KryoSerializer, importing
sqlContext.createSchemaRDD causes multiple spark applications creations )
> When spark.serializer is set as org.apache.spark.serializer.KryoSerializer,
> importing a method causes multiple spark applications creations
> ---------------------------------------------------------------------------------------------------------------------------------------------
>
> Key: SPARK-2610
> URL: https://issues.apache.org/jira/browse/SPARK-2610
> Project: Spark
> Issue Type: Bug
> Affects Versions: 1.0.1
> Reporter: Yin Huai
>
> To reproduce, set
> {code}
> spark.serializer org.apache.spark.serializer.KryoSerializer
> {code}
> in conf/spark-defaults.conf and launch a spark shell.
> Then, execute
> {code}
> val sqlContext = new org.apache.spark.sql.SQLContext(sc)
> import sqlContext.createSchemaRDD
> case class Person(name: String, age: Int)
> val people =
> sc.textFile("examples/src/main/resources/people.txt").map(_.split(",")).map(p
> => Person(p(0), p(1).trim.toInt))
> people.collect
> {code}
> There is no extra spark application creations if you remove
> {code}
> import sqlContext.createSchemaRDD
> {code}
> Our current branch 1.0 also has this issue.
--
This message was sent by Atlassian JIRA
(v6.2#6252)