Oz Ben-Ami created SPARK-22575: ---------------------------------- Summary: Making Spark Thrift Server clean up its cache Key: SPARK-22575 URL: https://issues.apache.org/jira/browse/SPARK-22575 Project: Spark Issue Type: Improvement Components: Block Manager, SQL Affects Versions: 2.2.0 Reporter: Oz Ben-Ami Priority: Minor
Currently, Spark Thrift Server accumulates data in its appcache, even for old queries. This fills up the disk (using over 100GB per worker node) within days, and the only way to clear it is to restart the Thrift Server application. Even deleting the files directly isn't a solution, as Spark then complains about FileNotFound. I asked about this on [Stack Overflow|https://stackoverflow.com/questions/46893123/how-can-i-make-spark-thrift-server-clean-up-its-cache] a few weeks ago, but it does not seem to be currently doable by configuration. Am I missing some configuration option, or some other factor here? Otherwise, can anyone point me to the code that handles this, so maybe I can try my hand at a fix? Thanks! -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org