What error? On Mon, Jan 18, 2016 at 9:01 AM, Simon Hafner <reactorm...@gmail.com> wrote:
> And for deserializing, > `sqlContext.read.parquet("path/to/parquet").as[T]` and catch the > error? > > 2016-01-14 3:43 GMT+08:00 Michael Armbrust <mich...@databricks.com>: > > Yeah, thats the best way for now (note the conversion is purely logical > so > > there is no cost of calling toDF()). We'll likely be combining the > classes > > in Spark 2.0 to remove this awkwardness. > > > > On Tue, Jan 12, 2016 at 11:20 PM, Simon Hafner <reactorm...@gmail.com> > > wrote: > >> > >> What's the proper way to write DataSets to disk? Convert them to a > >> DataFrame and use the writers there? > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > >> For additional commands, e-mail: user-h...@spark.apache.org > >> > > >