Hi,

Thank you for your reply. It surely going to help.

Regards,
Abhishek Dubey

From: Cheng, Hao [mailto:hao.ch...@intel.com]
Sent: Monday, March 02, 2015 6:52 PM
To: Abhishek Dubey; user@spark.apache.org
Subject: RE: Performance tuning in Spark SQL.


This is actually a quite open question, from my understanding, there're 
probably ways to tune like:



*        SQL Configurations like:



Configuration Key


Default Value


spark.sql.autoBroadcastJoinThreshold


10 * 1024 * 1024


spark.sql.defaultSizeInBytes


10 * 1024 * 1024 + 1


spark.sql.planner.externalSort


false


spark.sql.shuffle.partitions


200


spark.sql.codegen


false




*        Spark Cluster / Application Configuration (Memory, GC etc. Spark Core 
Number etc.)

*        Try using the Cached tables / Parquet Files as the storage.

*        "EXPLAIN [EXTENDED] query" is your best friend to tuning your SQL 
itself.

*        ...



And, a real use case scenario probably be more helpful in answering your 
question.



-----Original Message-----
From: dubey_a [mailto:abhishek.du...@xoriant.com]
Sent: Monday, March 2, 2015 6:02 PM
To: user@spark.apache.org<mailto:user@spark.apache.org>
Subject: Performance tuning in Spark SQL.



What are the ways to tune query performance in Spark SQL?







--

View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Performance-tuning-in-Spark-SQL-tp21871.html

Sent from the Apache Spark User List mailing list archive at Nabble.com.



---------------------------------------------------------------------

To unsubscribe, e-mail: 
user-unsubscr...@spark.apache.org<mailto:user-unsubscr...@spark.apache.org> For 
additional commands, e-mail: 
user-h...@spark.apache.org<mailto:user-h...@spark.apache.org>


Reply via email to