Hi All,

I have some questions about OFF_HEAP Caching. In Spark 1.X when we use 
rdd.persist(StorageLevel.OFF_HEAP),that means rdd caching in Tachyon(Alluxio). 
However,in Spark 2.X,we can directly use OFF_HEAP  For Caching

(https://issues.apache.org/jira/browse/SPARK-13992?jql=project%20%3D%20SPARK%20AND%20text%20~%20%22off-heap%20caching%22).
 I am confuse about this and I have follow questions:

1.In Spark 2.X, how should we use Tachyon for caching?

2.Is there any reason that must change in this way(I mean use off_heap directly 
instead of using Tachyon)

Thanks a lot!


Reply via email to