Have you tried to set the following? spark.worker.cleanup.enabled=true spark.worker.cleanup.appDataTtl=<seconds>”
On Thu, May 7, 2015 at 2:39 AM, Taeyun Kim <taeyun....@innowireless.com> wrote: > Hi, > > > > After a spark program completes, there are 3 temporary directories remain > in the temp directory. > > The file names are like this: spark-2e389487-40cc-4a82-a5c7-353c0feefbb7 > > > > And the Spark program runs on Windows, a snappy DLL file also remains in > the temp directory. > > The file name is like this: > snappy-1.0.4.1-6e117df4-97b6-4d69-bf9d-71c4a627940c-snappyjava > > > > They are created every time the Spark program runs. So the number of files > and directories keeps growing. > > > > How can let them be deleted? > > > > Spark version is 1.3.1 with Hadoop 2.6. > > > > Thanks. > > > > >