Hi,
 I think you are using the local model of Spark. There are mainly four models, which are local, standalone,  yarn and Mesos. Also, "blocks" is relative to hdfs, "partitions" is relative to spark.

liangyihuai

---Original---
From: "Jacek Laskowski "<ja...@japila.pl>
Date: 2017/2/25 02:45:20
To: "prithish"<prith...@gmail.com>;
Cc: "user"<user@spark.apache.org>;
Subject: Re: RDD blocks on Spark Driver

Hi, 

Guess you're use local mode which has only one executor called driver. Is my guessing correct? 

Jacek

On 23 Feb 2017 2:03 a.m., <prith...@gmail.com> wrote:
Hello,

Had a question. When I look at the executors tab in Spark UI, I notice that some RDD blocks are assigned to the driver as well. Can someone please tell me why?

Thanks for the help.

Reply via email to