Thanks for the response.

But no this does not answer the question.

The question was: Is there a way (via some API call) to query the number
and type of daemons currently running in the Spark cluster.

Regards


On Sun, Apr 26, 2015 at 10:12 AM, ayan guha <guha.a...@gmail.com> wrote:

> In my limited understanding, there must be single   "leader" master  in
> the cluster. If there are multiple leaders, it will lead to unstable
> cluster as each masters will keep scheduling independently. You should use
> zookeeper for HA, so that standby masters can vote to find new leader if
> the primary goes down.
>
> Now, you can still have multiple masters running as leaders but
> conceptually they should be thought as different clusters.
>
> Regarding workers, they should follow their master.
>
> Not sure if this answers your question, as I am sure you have read the
> documentation thoroughly.
>
> Best
> Ayan
>
> On Sun, Apr 26, 2015 at 6:31 PM, James King <jakwebin...@gmail.com> wrote:
>
>> If I have 5 nodes and I wish to maintain 1 Master and 2 Workers on each
>> node, so in total I will have 5 master and 10 Workers.
>>
>> Now to maintain that setup I would like to query spark regarding the
>> number Masters and Workers that are currently available using API calls and
>> then take some appropriate action based on the information I get back, like
>> restart a dead Master or Worker.
>>
>> Is this possible? does Spark provide such API?
>>
>
>
>
> --
> Best Regards,
> Ayan Guha
>

Reply via email to