The performance has more to do with the particular format you are using, not where the metadata is coming from. Even hive tables are read from files HDFS usually.
You probably should use HiveContext as its query language is more powerful than SQLContext. Also, parquet is usually the faster data format for Spark SQL. On Tue, Mar 17, 2015 at 3:41 AM, 李铖 <lidali...@gmail.com> wrote: > Hi,everybody. > > I am new in spark. Now I want to do interactive sql query using spark sql. > spark sql can run under hive or loading files from hdfs. > > Which is better or faster? > > Thanks. >