Hosur Narahari created SPARK-20093:
--------------------------------------

             Summary: Exception when Joining dataframe with another dataframe 
generated by applying groupBy transformation on original one
                 Key: SPARK-20093
                 URL: https://issues.apache.org/jira/browse/SPARK-20093
             Project: Spark
          Issue Type: Bug
          Components: SQL
    Affects Versions: 2.1.0, 2.0.2, 2.0.1, 2.0.0, 2.2.0
            Reporter: Hosur Narahari


When we generate a dataframe by doing grouping, and perform join on original 
dataframe with aggregate column, we get AnalysisException. Below I've attached 
a piece of code and resulting exception to reproduce.

Code:

import org.apache.spark.sql.SparkSession


object App {

  lazy val spark = 
SparkSession.builder.appName("Test").master("local").getOrCreate

  def main(args: Array[String]): Unit = {
    test1
  }

  private def test1 {
    import org.apache.spark.sql.functions._
    val df = spark.createDataFrame(Seq(("M",172,60), ("M", 170, 60), ("F", 155, 
56), ("M", 160, 55), ("F", 150, 53))).toDF("gender", "height", "weight")
    val groupDF = df.groupBy("gender").agg(min("height").as("height"))
    groupDF.show()
    val out = groupDF.join(df, groupDF("height") <=> 
df("height")).select(df("gender"), df("height"), df("weight"))
    out.show
  }
}

When I ran above code, I got below exception:

Exception in thread "main" org.apache.spark.sql.AnalysisException: resolved 
attribute(s) height#8 missing from 
height#19,height#30,gender#29,weight#31,gender#7 in operator !Join Inner, 
(height#19 <=> height#8);;
!Join Inner, (height#19 <=> height#8)
:- Aggregate [gender#7], [gender#7, min(height#8) AS height#19]
:  +- Project [_1#0 AS gender#7, _2#1 AS height#8, _3#2 AS weight#9]
:     +- LocalRelation [_1#0, _2#1, _3#2]
+- Project [_1#0 AS gender#29, _2#1 AS height#30, _3#2 AS weight#31]
   +- LocalRelation [_1#0, _2#1, _3#2]

        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:39)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:90)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:342)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:78)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:127)
        at 
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.checkAnalysis(CheckAnalysis.scala:78)
        at 
org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:90)
        at 
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:53)
        at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67)
        at 
org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$withPlan(Dataset.scala:2831)
        at org.apache.spark.sql.Dataset.join(Dataset.scala:843)
        at org.apache.spark.sql.Dataset.join(Dataset.scala:807)
        at App$.test1(App.scala:17)
        at App$.main(App.scala:9)
        at App.main(App.scala)

Please someone look into it.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to