isn’t an obvious mapping from Set[T] to
> any SQL types. Currently we have complex types like array, map, and struct,
> which are inherited from Hive. In your case, I’d transform the Set[T]
> into a Seq[T] first, then Spark SQL can map it to an array.
>
> Cheng
>
> On 1/28/15 7:
Hello,
We are trying to insert a case class in Parquet using SparkSql. When i'm
creating the SchemaRDD, that include a Set, i have the following exception:
sqc.createSchemaRDD(r)
scala.MatchError: Set[(scala.Int, scala.Int)] (of class
scala.reflect.internal.Types$TypeRef$$anon$1)
at
org.apache.sp
Hi i'm reading the O´really´s book Learning Spark and i have a doubt, the
accumulator's fault tolerance still only happening in the actions
operations? this behaviour is also expected if we use accumulables?
Thank in advance
Jorge López-Malla Matute
Big Data Developer
Avenida de Europa, 26. Áti