And for deserializing, `sqlContext.read.parquet("path/to/parquet").as[T]` and catch the error?
2016-01-14 3:43 GMT+08:00 Michael Armbrust <mich...@databricks.com>: > Yeah, thats the best way for now (note the conversion is purely logical so > there is no cost of calling toDF()). We'll likely be combining the classes > in Spark 2.0 to remove this awkwardness. > > On Tue, Jan 12, 2016 at 11:20 PM, Simon Hafner <reactorm...@gmail.com> > wrote: >> >> What's the proper way to write DataSets to disk? Convert them to a >> DataFrame and use the writers there? >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org