Is your query very complicated? Could you provide the output of `explain`
your query that consumes an excessive amount of memory? If this is a small
query, there may be a bug that leaks memory in SQLListener.

Best Regards,
Shixiong Zhu

2015-10-13 11:44 GMT+08:00 Nicholas Pritchard <
nicholas.pritch...@falkonry.com>:

> As an update, I did try disabling the ui with "spark.ui.enabled=false",
> but the JobListener and SQLListener still consume a lot of memory, leading
> to OOM error. Has anyone encountered this before? Is the only solution just
> to increase the driver heap size?
>
> Thanks,
> Nick
>
> On Mon, Oct 12, 2015 at 8:42 PM, Nicholas Pritchard <
> nicholas.pritch...@falkonry.com> wrote:
>
>> I set those configurations by passing to spark-submit script:
>> "bin/spark-submit --conf spark.ui.retainedJobs=20 ...". I have verified
>> that these configurations are being passed correctly because they are
>> listed in the environments tab and also by counting the number of
>> job/stages that are listed. The "spark.sql.ui.retainedExecutions=0" only
>> applies to the number of "completed" executions; there will always be a
>> "running" execution. For some reason, I have one execution that consumes an
>> excessive amount of memory.
>>
>> Actually, I am not interested in the SQL UI, as I find the Job/Stages UI
>> to have sufficient information. I am also using Spark Standalone cluster
>> manager so have not had to use the history server.
>>
>>
>> On Mon, Oct 12, 2015 at 8:17 PM, Shixiong Zhu <zsxw...@gmail.com> wrote:
>>
>>> Could you show how did you set the configurations? You need to set these
>>> configurations before creating SparkContext and SQLContext.
>>>
>>> Moreover, the history sever doesn't support SQL UI. So
>>> "spark.eventLog.enabled=true" doesn't work now.
>>>
>>> Best Regards,
>>> Shixiong Zhu
>>>
>>> 2015-10-13 2:01 GMT+08:00 pnpritchard <nicholas.pritch...@falkonry.com>:
>>>
>>>> Hi,
>>>>
>>>> In my application, the Spark UI is consuming a lot of memory,
>>>> especially the
>>>> SQL tab. I have set the following configurations to reduce the memory
>>>> consumption:
>>>> - spark.ui.retainedJobs=20
>>>> - spark.ui.retainedStages=40
>>>> - spark.sql.ui.retainedExecutions=0
>>>>
>>>> However, I still get OOM errors in the driver process with the default
>>>> 1GB
>>>> heap size. The following link is a screen shot of a heap dump report,
>>>> showing the SQLListener instance having a retained size of 600MB.
>>>>
>>>> https://cloud.githubusercontent.com/assets/5124612/10404379/20fbdcfc-6e87-11e5-9415-27e25193a25c.png
>>>>
>>>> Rather than just increasing the allotted heap size, does anyone have any
>>>> other ideas? Is it possible to disable the SQL tab specifically? I also
>>>> thought about serving the UI from disk rather than memory with
>>>> "spark.eventLog.enabled=true" and "spark.ui.enabled=false". Has anyone
>>>> tried
>>>> this before?
>>>>
>>>> Thanks,
>>>> Nick
>>>>
>>>>
>>>>
>>>> --
>>>> View this message in context:
>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Spark-UI-consuming-lots-of-memory-tp25033.html
>>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>>>> For additional commands, e-mail: user-h...@spark.apache.org
>>>>
>>>>
>>>
>>
>

Reply via email to