*Subject:* Re: Question about Serialization in Storage Level
From the docs,
https://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence
:
*Storage Level*
*Meaning*
MEMORY_ONLY
Store RDD as deserialized Java objects in the JVM. If the RDD does not fit
in memory
Hi there,
This question may seem to be kind of naïve, but what's the difference between
MEMORY_AND_DISK and MEMORY_AND_DISK_SER?
If I call rdd.persist(StorageLevel.MEMORY_AND_DISK), the BlockManager won't
serialize the rdd?
Thanks,
Zhipeng
From the docs,
https://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence:
Storage LevelMeaningMEMORY_ONLYStore RDD as deserialized Java objects in
the JVM. If the RDD does not fit in memory, some partitions will not be
cached and will be recomputed on the fly each time they're