The occasional type error if the casting goes wrong for whatever reason.
2016-01-19 1:22 GMT+08:00 Michael Armbrust :
> What error?
>
> On Mon, Jan 18, 2016 at 9:01 AM, Simon Hafner wrote:
>>
>> And for deserializing,
>> `sqlContext.read.parquet("path/to/parquet").as[T]` and catch the
>> error?
>
What error?
On Mon, Jan 18, 2016 at 9:01 AM, Simon Hafner wrote:
> And for deserializing,
> `sqlContext.read.parquet("path/to/parquet").as[T]` and catch the
> error?
>
> 2016-01-14 3:43 GMT+08:00 Michael Armbrust :
> > Yeah, thats the best way for now (note the conversion is purely logical
> so
And for deserializing,
`sqlContext.read.parquet("path/to/parquet").as[T]` and catch the
error?
2016-01-14 3:43 GMT+08:00 Michael Armbrust :
> Yeah, thats the best way for now (note the conversion is purely logical so
> there is no cost of calling toDF()). We'll likely be combining the classes
> i
Yeah, thats the best way for now (note the conversion is purely logical so
there is no cost of calling toDF()). We'll likely be combining the classes
in Spark 2.0 to remove this awkwardness.
On Tue, Jan 12, 2016 at 11:20 PM, Simon Hafner
wrote:
> What's the proper way to write DataSets to disk?
What's the proper way to write DataSets to disk? Convert them to a
DataFrame and use the writers there?
-
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org