It's not a Spark Streaming app, so sorry I'm not sure of the answer to that. I would assume it should work.
On Fri, May 15, 2015 at 2:22 PM, Evo Eftimov <evo.efti...@isecc.com> wrote: > Ok thanks a lot for clarifying that – btw was your application a Spark > Streaming App – I am also looking for confirmation that FAIR scheduling is > supported for Spark Streaming Apps > > > > *From:* Richard Marscher [mailto:rmarsc...@localytics.com] > *Sent:* Friday, May 15, 2015 7:20 PM > *To:* Evo Eftimov > *Cc:* Tathagata Das; user > *Subject:* Re: Spark Fair Scheduler for Spark Streaming - 1.2 and beyond > > > > The doc is a bit confusing IMO, but at least for my application I had to > use a fair pool configuration to get my stages to be scheduled with FAIR. > > > > On Fri, May 15, 2015 at 2:13 PM, Evo Eftimov <evo.efti...@isecc.com> > wrote: > > No pools for the moment – for each of the apps using the straightforward > way with the spark conf param for scheduling = FAIR > > > > Spark is running in a Standalone Mode > > > > Are you saying that Configuring Pools is mandatory to get the FAIR > scheduling working – from the docs it seemed optional to me > > > > *From:* Tathagata Das [mailto:t...@databricks.com] > *Sent:* Friday, May 15, 2015 6:45 PM > *To:* Evo Eftimov > *Cc:* user > *Subject:* Re: Spark Fair Scheduler for Spark Streaming - 1.2 and beyond > > > > How are you configuring the fair scheduler pools? > > > > On Fri, May 15, 2015 at 8:33 AM, Evo Eftimov <evo.efti...@isecc.com> > wrote: > > I have run / submitted a few Spark Streaming apps configured with Fair > scheduling on Spark Streaming 1.2.0, however they still run in a FIFO mode. > Is FAIR scheduling supported at all for Spark Streaming apps and from what > release / version - e.g. 1.3.1 > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Fair-Scheduler-for-Spark-Streaming-1-2-and-beyond-tp22902.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > > > > >