I will give that a try. Thanks! On Fri, Dec 16, 2016 at 12:45 AM, Michael Armbrust <mich...@databricks.com> wrote:
> I would have sworn there was a ticket, but I can't find it. So here you > go: https://issues.apache.org/jira/browse/SPARK-18891 > > A work around until that is fixed would be for you to manually specify the > kryo > encoder > <http://spark.apache.org/docs/2.0.2/api/java/org/apache/spark/sql/Encoders.html#kryo(scala.reflect.ClassTag)> > . > > On Thu, Dec 15, 2016 at 8:18 AM, Jakub Dubovsky < > spark.dubovsky.ja...@gmail.com> wrote: > >> Hey, >> >> I want to ask whether there is any roadmap/plan for adding Encoders for >> further types in next releases of Spark. Here is a list >> <http://spark.apache.org/docs/latest/sql-programming-guide.html#data-types> >> of >> currently supported types. We would like to use Datasets with our >> internally defined case classes containing >> scala.collection.immutable.List(s). >> This does not work now because these lists are converted to ArrayType >> (Seq). This then fails a constructor lookup because of seq-is-not-a-list >> error... >> >> This means that for now we are stuck with using RDDs. >> >> Thanks for any insights! >> >> Jakub Dubovsky >> >> >