So if I do a spark action, say, collect, will I be able to see the result
on my local console? Or would it be only available only on the cluster
master?


On Mon, Feb 17, 2014 at 9:50 AM, purav aggarwal
<puravaggarwal...@gmail.com>wrote:

> Your local machine simply submits your job (in the form of jar) to the
> cluster.
> The master node is where the SparkContext object is created, a DAG of your
> job is formed and tasks (stages) are assigned to different workers - which
> are not aware of anything but computation of task being assigned.
>
>
> On Mon, Feb 17, 2014 at 10:07 PM, David Thomas <dt5434...@gmail.com>wrote:
>
>> Where is the SparkContext object created then? On my local machine or on
>> the master node in the cluster?
>>
>>
>> On Mon, Feb 17, 2014 at 4:17 AM, Nhan Vu Lam Chi <nhani...@adatao.com>wrote:
>>
>>> Your local app will be called "driver program", which creates jobs and
>>> submits them to the cluster for running.
>>>
>>>
>>> On Mon, Feb 17, 2014 at 9:19 AM, David Thomas <dt5434...@gmail.com>wrote:
>>>
>>>> From 
>>>> docs<https://spark.incubator.apache.org/docs/latest/spark-standalone.html>
>>>> :
>>>>
>>>>
>>>> *Connecting an Application to the ClusterTo run an application on the
>>>> Spark cluster, simply pass the spark://IP:PORT URL of the master as to the
>>>> SparkContext constructor.*
>>>>
>>>> Could someone enlighten me on what happens if I run the app, from say,
>>>> Eclipse on my local machine, but use the url of the master node which is on
>>>> cloud. What role does my local JVM play then?
>>>>
>>>
>>>
>>
>

Reply via email to