In addition, you cannot turn off JobListener and SQLListener now... Best Regards, Shixiong Zhu
2015-10-13 11:59 GMT+08:00 Shixiong Zhu <zsxw...@gmail.com>: > Is your query very complicated? Could you provide the output of `explain` > your query that consumes an excessive amount of memory? If this is a small > query, there may be a bug that leaks memory in SQLListener. > > Best Regards, > Shixiong Zhu > > 2015-10-13 11:44 GMT+08:00 Nicholas Pritchard < > nicholas.pritch...@falkonry.com>: > >> As an update, I did try disabling the ui with "spark.ui.enabled=false", >> but the JobListener and SQLListener still consume a lot of memory, leading >> to OOM error. Has anyone encountered this before? Is the only solution just >> to increase the driver heap size? >> >> Thanks, >> Nick >> >> On Mon, Oct 12, 2015 at 8:42 PM, Nicholas Pritchard < >> nicholas.pritch...@falkonry.com> wrote: >> >>> I set those configurations by passing to spark-submit script: >>> "bin/spark-submit --conf spark.ui.retainedJobs=20 ...". I have verified >>> that these configurations are being passed correctly because they are >>> listed in the environments tab and also by counting the number of >>> job/stages that are listed. The "spark.sql.ui.retainedExecutions=0" >>> only applies to the number of "completed" executions; there will always be >>> a "running" execution. For some reason, I have one execution that consumes >>> an excessive amount of memory. >>> >>> Actually, I am not interested in the SQL UI, as I find the Job/Stages UI >>> to have sufficient information. I am also using Spark Standalone cluster >>> manager so have not had to use the history server. >>> >>> >>> On Mon, Oct 12, 2015 at 8:17 PM, Shixiong Zhu <zsxw...@gmail.com> wrote: >>> >>>> Could you show how did you set the configurations? You need to set >>>> these configurations before creating SparkContext and SQLContext. >>>> >>>> Moreover, the history sever doesn't support SQL UI. So >>>> "spark.eventLog.enabled=true" doesn't work now. >>>> >>>> Best Regards, >>>> Shixiong Zhu >>>> >>>> 2015-10-13 2:01 GMT+08:00 pnpritchard <nicholas.pritch...@falkonry.com> >>>> : >>>> >>>>> Hi, >>>>> >>>>> In my application, the Spark UI is consuming a lot of memory, >>>>> especially the >>>>> SQL tab. I have set the following configurations to reduce the memory >>>>> consumption: >>>>> - spark.ui.retainedJobs=20 >>>>> - spark.ui.retainedStages=40 >>>>> - spark.sql.ui.retainedExecutions=0 >>>>> >>>>> However, I still get OOM errors in the driver process with the default >>>>> 1GB >>>>> heap size. The following link is a screen shot of a heap dump report, >>>>> showing the SQLListener instance having a retained size of 600MB. >>>>> >>>>> https://cloud.githubusercontent.com/assets/5124612/10404379/20fbdcfc-6e87-11e5-9415-27e25193a25c.png >>>>> >>>>> Rather than just increasing the allotted heap size, does anyone have >>>>> any >>>>> other ideas? Is it possible to disable the SQL tab specifically? I also >>>>> thought about serving the UI from disk rather than memory with >>>>> "spark.eventLog.enabled=true" and "spark.ui.enabled=false". Has anyone >>>>> tried >>>>> this before? >>>>> >>>>> Thanks, >>>>> Nick >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-UI-consuming-lots-of-memory-tp25033.html >>>>> Sent from the Apache Spark User List mailing list archive at >>>>> Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>>>> For additional commands, e-mail: user-h...@spark.apache.org >>>>> >>>>> >>>> >>> >> >