Hi, Thank you for your reply. It surely going to help.
Regards, Abhishek Dubey From: Cheng, Hao [mailto:hao.ch...@intel.com] Sent: Monday, March 02, 2015 6:52 PM To: Abhishek Dubey; user@spark.apache.org Subject: RE: Performance tuning in Spark SQL. This is actually a quite open question, from my understanding, there're probably ways to tune like: * SQL Configurations like: Configuration Key Default Value spark.sql.autoBroadcastJoinThreshold 10 * 1024 * 1024 spark.sql.defaultSizeInBytes 10 * 1024 * 1024 + 1 spark.sql.planner.externalSort false spark.sql.shuffle.partitions 200 spark.sql.codegen false * Spark Cluster / Application Configuration (Memory, GC etc. Spark Core Number etc.) * Try using the Cached tables / Parquet Files as the storage. * "EXPLAIN [EXTENDED] query" is your best friend to tuning your SQL itself. * ... And, a real use case scenario probably be more helpful in answering your question. -----Original Message----- From: dubey_a [mailto:abhishek.du...@xoriant.com] Sent: Monday, March 2, 2015 6:02 PM To: user@spark.apache.org<mailto:user@spark.apache.org> Subject: Performance tuning in Spark SQL. What are the ways to tune query performance in Spark SQL? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Performance-tuning-in-Spark-SQL-tp21871.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org<mailto:user-unsubscr...@spark.apache.org> For additional commands, e-mail: user-h...@spark.apache.org<mailto:user-h...@spark.apache.org>