The fair scheduler doesn't have anything to do with reallocating resource
across Applications.

https://spark.apache.org/docs/latest/job-scheduling.html#scheduling-across-applications
https://spark.apache.org/docs/latest/job-scheduling.html#scheduling-within-an-application

On Thu, Jul 20, 2017 at 2:02 PM, Gokula Krishnan D <email2...@gmail.com>
wrote:

> Mark, Thanks for the response.
>
> Let me rephrase my statements.
>
> "I am submitting a Spark application(*Application*#A) with scheduler.mode
> as FAIR and dynamicallocation=true and it got all the available executors.
>
> In the meantime, submitting another Spark Application (*Application* # B)
> with the scheduler.mode as FAIR and dynamicallocation=true but it got only
> one executor. "
>
> Thanks & Regards,
> Gokula Krishnan* (Gokul)*
>
> On Thu, Jul 20, 2017 at 4:56 PM, Mark Hamstra <m...@clearstorydata.com>
> wrote:
>
>> First, Executors are not allocated to Jobs, but rather to Applications.
>> If you run multiple Jobs within a single Application, then each of the
>> Tasks associated with Stages of those Jobs has the potential to run on any
>> of the Application's Executors. Second, once a Task starts running on an
>> Executor, it has to complete before another Task can be scheduled using the
>> prior Task's resources -- the fair scheduler is not preemptive of running
>> Tasks.
>>
>> On Thu, Jul 20, 2017 at 1:45 PM, Gokula Krishnan D <email2...@gmail.com>
>> wrote:
>>
>>> Hello All,
>>>
>>> We are having cluster with 50 Executors each with 4 Cores so can avail
>>> max. 200 Executors.
>>>
>>> I am submitting a Spark application(JOB A) with scheduler.mode as FAIR
>>> and dynamicallocation=true and it got all the available executors.
>>>
>>> In the meantime, submitting another Spark Application (JOB B) with the
>>> scheduler.mode as FAIR and dynamicallocation=true but it got only one
>>> executor.
>>>
>>> Normally this situation occurs when any of the JOB runs with the
>>> Scheduler.mode= FIFO.
>>>
>>> 1) Have your ever faced this issue if so how to overcome this?.
>>>
>>> I was in the impression that as soon as I submit the JOB B the Spark
>>> Scheduler should distribute/release few resources from the JOB A and share
>>> it with the JOB A in the Round Robin fashion?.
>>>
>>> Appreciate your response !!!.
>>>
>>>
>>> Thanks & Regards,
>>> Gokula Krishnan* (Gokul)*
>>>
>>
>>
>

Reply via email to