[ https://issues.apache.org/jira/browse/SPARK-6156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14346797#comment-14346797 ]
Apache Spark commented on SPARK-6156: ------------------------------------- User 'suyanNone' has created a pull request for this issue: https://github.com/apache/spark/pull/4886 > Refine Put Memory_And_Disk block > -------------------------------- > > Key: SPARK-6156 > URL: https://issues.apache.org/jira/browse/SPARK-6156 > Project: Spark > Issue Type: Bug > Components: Spark Core > Reporter: SuYan > Fix For: 1.3.0 > > > Current code: > Now we want to cache a Memory_and_Disk level block > 1. Try to put in memory and unroll unsuccessful. then reserved unroll memory > because we got a iterator from an unroll Array > 2. Then put into disk. > 3. Get value from get(blockId), and iterator from that value. > And there have some doubts: > for process 3: > get value use blockManager.get(blockId) > in get(blockId) method, it will first getLocal() and then getRemote() in > logical, I think is more suitable for use getLocal(blockId) to instead > get(blockId), I dont know why we call get(blockId), have other considerations? > and in doGetLocal method, if it get value success from disk, it will try to > put in memory again... I think, that a redundant... we just try to put in > memory and failed to unroll in process 1. > may this pull request is useless, because may have some reasons to do that? -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org