If i understand correctly , the above document creates pool for priority
which is static in nature and has to be defined before submitting the job .
.in my scenario each generated task can have different priority.

Thanks,
Abhi


On Mon, Mar 16, 2015 at 9:48 PM, twinkle sachdeva <
twinkle.sachd...@gmail.com> wrote:

> Hi,
>
> Maybe this is what you are looking for :
> http://spark.apache.org/docs/1.2.0/job-scheduling.html#fair-scheduler-pools
>
> Thanks,
>
> On Mon, Mar 16, 2015 at 8:15 PM, abhi <abhishek...@gmail.com> wrote:
>
>> Hi
>> Current all the jobs in spark gets submitted using queue . i have a
>> requirement where submitted job will generate another set of jobs with some
>> priority , which should again be submitted to spark cluster based on
>> priority ? Means job with higher priority should be executed first,    Is
>> it feasible  ?
>>
>> Any help is appreciated ?
>>
>> Thanks,
>> Abhi
>>
>>
>
>

Reply via email to