The implicit conversion function mentioned by Hao is createSchemaRDD in
SQLContext/HiveContext.
You can import it by doing
val sqlContext = new org.apache.spark.sql.SQLContext(sc)
// Or new org.apache.spark.sql.hive.HiveContext(sc) for HiveContext
import sqlContext.createSchemaRDD
On Wed, Oct 22, 2014 at 8:03 AM, Cheng, Hao hao.ch...@intel.com wrote:
You needn’t do anything, the implicit conversion should do this for you.
https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala#L103
https://github.com/apache/spark/blob/2ac40da3f9fa6d45a59bb45b41606f1931ac5e81/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ScalaReflection.scala#L35
Just be sure you import the right implicit conversion function.
*From:* Dai, Kevin [mailto:yun...@ebay.com]
*Sent:* Wednesday, October 22, 2014 4:17 PM
*To:* user@spark.apache.org
*Subject:* SchemaRDD Convert
Hi, ALL
I have a RDD of case class T and T contains several primitive types and a
Map.
How can I convert this to a SchemaRDD?
Best Regards,
Kevin.