Thank you Jey, That is a nice introduction but it is a may be to old (AUG 21ST, 2013)
"Note: If you keep the schema flat (without nesting), the Parquet files you create can be read by systems like Shark and Impala. These systems allow you to query Parquet files as tables using SQL-like syntax. The Parquet files created by this sample application could easily be queried using Shark for example." But in this post (http://apache-spark-user-list.1001560.n3.nabble.com/SparkSQL-Nested-CaseClass-Parquet-failure-td8377.html) I found this: Nested parquet is not supported in 1.0, but is part of the upcoming 1.0.1 release. So the question now is, can I use it in the benefit way of nested parquet files to find fast with sql or do I have to write a special map/reduce job to transform and find my data? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Is-it-possible-to-use-Parquet-with-Dremel-encoding-tp15186p15234.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org