Do you know what version of spark you are running with?
On Thu, Dec 3, 2015 at 12:52 AM -0800, "Kevin (Sangwoo) Kim" <kevin...@apache.org> wrote: Do you use broadcast variables? I've found many problems related to broadcast variables and not using it. (It's a Spark problem, rather than Zeppelin problem) For RDD's, no need to be manually unpersisted, it automatically does. 2015년 12월 3일 (목) 오후 5:28, Jakub Liska <liska.ja...@gmail.com>님이 작성: > Hi, > > no, just running it manually. I think I need to unpersist cached rdds and > destroy broadcast variables in the end, am I correct? Because it hasn't > crashed since then, the following runs are always a little slower though. > > On Thu, Dec 3, 2015 at 8:08 AM, Felix Cheung <felixcheun...@hotmail.com> > wrote: > >> How are you running jobs? Do you schedule a notebook to run from Zeppelin? >> >> ------------------------------ >> Date: Mon, 30 Nov 2015 12:42:16 +0100 >> Subject: Spark worker memory not freed up after zeppelin run finishes >> From: liska.ja...@gmail.com >> To: users@zeppelin.incubator.apache.org >> >> Hey, >> >> I'm connecting Zeppelin with a remote Spark standalone cluster (2 worker >> nodes) and I noticed that if I run a job from Zeppelin twice without >> restarting the Interpreter, it fails on OOME. After the Zeppelin jobs >> successfully finishes I can see all executor memory being allocated on >> workers and restarting Interpreter frees the memory... But if I don't do it >> it fails when running the task again. >> >> Any idea how to deal with this problem? Currently I have to always >> restart Interpreter between running spark jobs. >> >> Thanks Jakub >> > >