http://apache-spark-developers-list.1001551.n3.nabble.com/Job-priority-td10076.html#a10079
On Mon, Mar 16, 2015 at 10:26 PM, abhi <abhishek...@gmail.com> wrote: > If i understand correctly , the above document creates pool for priority > which is static in nature and has to be defined before submitting the job . > .in my scenario each generated task can have different priority. > > Thanks, > Abhi > > > On Mon, Mar 16, 2015 at 9:48 PM, twinkle sachdeva < > twinkle.sachd...@gmail.com> wrote: > >> Hi, >> >> Maybe this is what you are looking for : >> http://spark.apache.org/docs/1.2.0/job-scheduling.html#fair-scheduler-pools >> >> Thanks, >> >> On Mon, Mar 16, 2015 at 8:15 PM, abhi <abhishek...@gmail.com> wrote: >> >>> Hi >>> Current all the jobs in spark gets submitted using queue . i have a >>> requirement where submitted job will generate another set of jobs with some >>> priority , which should again be submitted to spark cluster based on >>> priority ? Means job with higher priority should be executed first, Is >>> it feasible ? >>> >>> Any help is appreciated ? >>> >>> Thanks, >>> Abhi >>> >>> >> >> >