On cluster with HDFS + Spark (in standalone deploy mode), there is a master node + 4 worker nodes. When a spark-shell connects to master, it creates 4 executor JVMs on each of the 4 worker nodes.
When the application reads a HDFS files and does computations in RDDs, what work gets done on master, worker, executor and driver ? Thanks,