Try this one: df.groupBy( when(expr("field1='foo'"),"field1").when(expr("field2='bar'"),"field2"))
On Sun, Jun 4, 2017 at 3:16 AM, Bryan Jeffrey <bryan.jeff...@gmail.com> wrote: > You should be able to project a new column that is your group column. Then > you can group on the projected column. > > Get Outlook for Android <https://aka.ms/ghei36> > > > > > On Sat, Jun 3, 2017 at 6:26 PM -0400, "upendra 1991" < > upendra1...@yahoo.com.invalid> wrote: > > Use a function >> >> Sent from Yahoo Mail on Android >> <https://overview.mail.yahoo.com/mobile/?.src=Android> >> >> On Sat, Jun 3, 2017 at 5:01 PM, kant kodali >> <kanth...@gmail.com> wrote: >> Hi All, >> >> Is there a way to do conditional group by in spark 2.1.1? other words, I >> want to do something like this >> >> if (field1 == "foo") { >> df.groupBy(field1) >> } else if (field2 == "bar") >> df.groupBy(field2) >> >> Thanks >> >>