The performance has more to do with the particular format you are using,
not where the metadata is coming from.   Even hive tables are read from
files HDFS usually.

You probably should use HiveContext as its query language is more powerful
than SQLContext.  Also, parquet is usually the faster data format for Spark
SQL.

On Tue, Mar 17, 2015 at 3:41 AM, 李铖 <lidali...@gmail.com> wrote:

> Hi,everybody.
>
> I am new in spark. Now I want to do interactive sql query using spark sql.
> spark sql can run under hive or loading files from hdfs.
>
> Which is better or faster?
>
> Thanks.
>

Reply via email to