Jorge Machado created SPARK-24401: ------------------------------------- Summary: Aggreate on Decimal Types does not work Key: SPARK-24401 URL: https://issues.apache.org/jira/browse/SPARK-24401 Project: Spark Issue Type: Bug Components: SQL Affects Versions: 2.3.0, 2.2.0 Reporter: Jorge Machado Attachments: testDF.parquet
Hi, I think I found a really ugly bug in spark when performing aggregations with Decimals To reproduce: {code:java} val df = spark.read.parquet("attached file") val first_agg = fact_df.groupBy("id1", "id2", "start_date").agg(mean("projection_factor").alias("projection_factor")) first_agg.show val second_agg = first_agg.groupBy("id1","id2").agg(max("projection_factor").alias("maxf"), min("projection_factor").alias("minf")) second_agg.show {code} First aggregation works fine the second aggregation seems to be summing instead of max value. I tried with spark 2.2.0 and 2.3.0 same problem. -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org