https://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.sql.hive.HiveContext
I'm getting org.apache.spark.sql.catalyst.analysis.NoSuchTableException from: val dataframe = hiveContext.table("other_db.mytable") Do I have to change current database to access it? Is it possible to do this? I'm guessing that the "database.table" syntax that I used in hiveContext.table is not recognized. I have no problems accessing tables in the database called "default". I can list tables in "other_db" with hiveContext.tableNames("other_db") Using Spark 1.4.0.