I would have sworn there was a ticket, but I can't find it. So here you go: https://issues.apache.org/jira/browse/SPARK-18891
A work around until that is fixed would be for you to manually specify the kryo encoder <http://spark.apache.org/docs/2.0.2/api/java/org/apache/spark/sql/Encoders.html#kryo(scala.reflect.ClassTag)> . On Thu, Dec 15, 2016 at 8:18 AM, Jakub Dubovsky < spark.dubovsky.ja...@gmail.com> wrote: > Hey, > > I want to ask whether there is any roadmap/plan for adding Encoders for > further types in next releases of Spark. Here is a list > <http://spark.apache.org/docs/latest/sql-programming-guide.html#data-types> of > currently supported types. We would like to use Datasets with our > internally defined case classes containing scala.collection.immutable.List(s). > This does not work now because these lists are converted to ArrayType > (Seq). This then fails a constructor lookup because of seq-is-not-a-list > error... > > This means that for now we are stuck with using RDDs. > > Thanks for any insights! > > Jakub Dubovsky > >