Yin Huai created SPARK-2610:
-------------------------------
Summary: When spark.serializer is set as
org.apache.spark.serializer.KryoSerializer, importing
sqlContext.createSchemaRDD causes multiple spark applications creations
Key: SPARK-2610
URL: https://issues.apache.org/jira/browse/SPARK-2610
Project: Spark
Issue Type: Bug
Affects Versions: 1.0.1
Reporter: Yin Huai
To reproduce, set
{code}
spark.serializer org.apache.spark.serializer.KryoSerializer
{code}
in conf/spark-defaults.conf and launch a spark shell.
Then, execute
{code}
val sqlContext = new org.apache.spark.sql.SQLContext(sc)
import sqlContext.createSchemaRDD
case class Person(name: String, age: Int)
val people =
sc.textFile("examples/src/main/resources/people.txt").map(_.split(",")).map(p
=> Person(p(0), p(1).trim.toInt))
people.collect
{code}
There is no extra spark application creations if you remove
{code}
import sqlContext.createSchemaRDD
{code}
Our current branch 1.0 also has this issue.
--
This message was sent by Atlassian JIRA
(v6.2#6252)