It will spill to disk if everything can’t be loaded in memory .
On Wed, 22 Jun 2022 at 5:58 PM, Sid <flinkbyhe...@gmail.com> wrote: > I have a 150TB CSV file. > > I have a total of 100 TB RAM and 100TB disk. So If I do something like this > > spark.read.option("header","true").csv(filepath).show(false) > > Will it lead to an OOM error since it doesn't have enough memory? or it > will spill data onto the disk and process it? > > Thanks, > Sid > -- Thanks Deepak www.bigdatabig.com www.keosha.net