It is called groupByKey now. Similar to joinWith, the schema produced by
relational joins and aggregations is different than what you would expect
when working with objects. So, when combining DataFrame+Dataset we renamed
these functions to make this distinction clearer.
On Sun, Apr 3, 2016 at
Hi,
(since 2.0.0-SNAPSHOT it's more for dev not user)
With today's master I'm getting the following:
scala> ds
res14: org.apache.spark.sql.Dataset[(String, Int)] = [_1: string, _2: int]
// WHY?!
scala> ds.groupBy(_._1)
:26: error: missing parameter type for expanded function
((x$1) => x$1._1)