[ https://issues.apache.org/jira/browse/SPARK-1042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14176708#comment-14176708 ]
Tal Sliwowicz commented on SPARK-1042: -------------------------------------- [~qqsun8819] I think the issue was resolved in 0.9.2. We are not experiencing it any more. Thanks! > spark cleans all java broadcast variables when it hits the spark.cleaner.ttl > ----------------------------------------------------------------------------- > > Key: SPARK-1042 > URL: https://issues.apache.org/jira/browse/SPARK-1042 > Project: Spark > Issue Type: Bug > Components: Java API, Spark Core > Affects Versions: 0.8.0, 0.8.1, 0.9.0 > Reporter: Tal Sliwowicz > Assignee: OuyangJin > Priority: Critical > Labels: memory_leak > > When setting spark.cleaner.ttl, spark performs the cleanup on time - but it > cleans all broadcast variables, not just the ones that are older than the > ttl. This creates an exception when the next mapPartitions runs because it > cannot find the broadcast variable, even when it was created immediately > before running the task. > Our temp workaround - not set the ttl and suffer from an ongoing memory leak > (forces a restart). > We are using JavaSparkContext and our broadcast variables are Java HashMaps. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org