Hi, If I have a table in Hive metastore saved as Parquet, and I want to use it in Spark. It seems Spark will use Hive's Parquet serde to load the actual data.
So is there any difference here? Will predicate pushdown, pruning and future Parquet optimizations in SparkSQL work for using Hive serde? Loading tables using parquetFile vs. loading tables from Hive metastore with Parquet serde Thanks, -- Jianshi Huang LinkedIn: jianshi Twitter: @jshuang Github & Blog: http://huangjs.github.com/