Yin Huai created SPARK-2610:
-------------------------------

             Summary: When spark.serializer is set as 
org.apache.spark.serializer.KryoSerializer, importing 
sqlContext.createSchemaRDD causes multiple spark applications creations  
                 Key: SPARK-2610
                 URL: https://issues.apache.org/jira/browse/SPARK-2610
             Project: Spark
          Issue Type: Bug
    Affects Versions: 1.0.1
            Reporter: Yin Huai


To reproduce, set
{code}
spark.serializer        org.apache.spark.serializer.KryoSerializer
{code}
in conf/spark-defaults.conf and launch a spark shell.
Then, execute
{code}
val sqlContext = new org.apache.spark.sql.SQLContext(sc)
import sqlContext.createSchemaRDD
case class Person(name: String, age: Int)
val people = 
sc.textFile("examples/src/main/resources/people.txt").map(_.split(",")).map(p 
=> Person(p(0), p(1).trim.toInt))
people.collect
{code}

There is no extra spark application creations if you remove
{code}
import sqlContext.createSchemaRDD
{code}

Our current branch 1.0 also has this issue. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to