Hi Tridib, I changed SQLContext to HiveContext and it started working. These are steps I used.
val sqlContext = new org.apache.spark.sql.hive.HiveContext(sc) val person = sqlContext.jsonFile("json/person.json") person.printSchema() person.registerTempTable("person") val address = sqlContext.jsonFile("json/address.json") address.printSchema() address.registerTempTable("address") sqlContext.cacheTable("person") sqlContext.cacheTable("address") val rs2 = sqlContext.sql("select p.id,p.name,a.city from person p join address a on (p.id = a.id)").collect.foreach(println) Rishi@InfoObjects *Pure-play Big Data Consulting* On Tue, Oct 21, 2014 at 5:47 AM, tridib <tridib.sama...@live.com> wrote: > val sqlContext = new org.apache.spark.sql.SQLContext(sc) > val personPath = "/hdd/spark/person.json" > val person = sqlContext.jsonFile(personPath) > person.printSchema() > person.registerTempTable("person") > val addressPath = "/hdd/spark/address.json" > val address = sqlContext.jsonFile(addressPath) > address.printSchema() > address.registerTempTable("address") > sqlContext.cacheTable("person") > sqlContext.cacheTable("address") > val rs2 = sqlContext.sql("SELECT p.id, p.name, a.city FROM person p, > address > a where p.id = a.id limit 10").collect.foreach(println) > > person.json > {"id:"1","name":"Mr. X"} > > address.json > {"city:"Earth","id":"1"} > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/spark-sql-join-sql-fails-after-sqlCtx-cacheTable-tp16893p16914.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >