Great stuff Tim. This definitely will make Mesos users life easier Sent from my iPad
On 2015-08-12, at 11:52, Haripriya Ayyalasomayajula <aharipriy...@gmail.com> wrote: > Thanks Tim, Jerry. > > On Wed, Aug 12, 2015 at 1:18 AM, Tim Chen <t...@mesosphere.io> wrote: > Yes the options are not that configurable yet but I think it's not hard to > change it. > > I have a patch out actually specifically able to configure amount of cpus per > executor in coarse grain mode, and hopefully merged next release. > > I think the open question now is for fine grain mode can we limit the number > of maximum concurrent executors, and I think we can definitely just add a new > option like spark.mesos.executor.max to cap it. > > I'll file a jira and hopefully to get this change in soon too. > > Tim > > > > On Tue, Aug 11, 2015 at 6:21 AM, Haripriya Ayyalasomayajula > <aharipriy...@gmail.com> wrote: > Spark evolved as an example framework for Mesos - thats how I know it. It is > surprising to see that the options provided by mesos in this case are less. > Tweaking the source code, haven't done it yet but I would love to see what > options could be there! > > On Tue, Aug 11, 2015 at 5:42 AM, Jerry Lam <chiling...@gmail.com> wrote: > My experience with Mesos + Spark is not great. I saw one executor with 30 CPU > and the other executor with 6. So I don't think you can easily configure it > without some tweaking at the source code. > > Sent from my iPad > > On 2015-08-11, at 2:38, Haripriya Ayyalasomayajula <aharipriy...@gmail.com> > wrote: > >> Hi Tim, >> >> Spark on Yarn allows us to do it using --num-executors and --executor_cores >> commandline arguments. I just got a chance to look at a similar spark user >> list mail, but no answer yet. So does mesos allow setting the number of >> executors and cores? Is there a default number it assumes? >> >> On Mon, Jan 5, 2015 at 5:07 PM, Tim Chen <t...@mesosphere.io> wrote: >> Forgot to hit reply-all. >> >> ---------- Forwarded message ---------- >> From: Tim Chen <t...@mesosphere.io> >> Date: Sun, Jan 4, 2015 at 10:46 PM >> Subject: Re: Controlling number of executors on Mesos vs YARN >> To: mvle <m...@us.ibm.com> >> >> >> Hi Mike, >> >> You're correct there is no such setting in for Mesos coarse grain mode, >> since the assumption is that each node is launched with one container and >> Spark is launching multiple tasks in that container. >> >> In fine-grain mode there isn't a setting like that, as it currently will >> launch an executor as long as it satisfies the minimum container resource >> requirement. >> >> I've created a JIRA earlier about capping the number of executors or better >> distribute the # of executors launched in each node. Since the decision of >> choosing what node to launch containers is all in the Spark scheduler side, >> it's very easy to modify it. >> >> Btw, what's the configuration to set the # of executors on YARN side? >> >> Thanks, >> >> Tim >> >> >> >> On Sun, Jan 4, 2015 at 9:37 PM, mvle <m...@us.ibm.com> wrote: >> I'm trying to compare the performance of Spark running on Mesos vs YARN. >> However, I am having problems being able to configure the Spark workload to >> run in a similar way on Mesos and YARN. >> >> When running Spark on YARN, you can specify the number of executors per >> node. So if I have a node with 4 CPUs, I can specify 6 executors on that >> node. When running Spark on Mesos, there doesn't seem to be an equivalent >> way to specify this. In Mesos, you can somewhat force this by specifying the >> number of CPU resources to be 6 when running the slave daemon. However, this >> seems to be a static configuration of the Mesos cluster rather something >> that can be configured in the Spark framework. >> >> So here is my question: >> >> For Spark on Mesos, am I correct that there is no way to control the number >> of executors per node (assuming an idle cluster)? For Spark on Mesos >> coarse-grained mode, there is a way to specify max_cores but that is still >> not equivalent to specifying the number of executors per node as when Spark >> is run on YARN. >> >> If I am correct, then it seems Spark might be at a disadvantage running on >> Mesos compared to YARN (since it lacks the fine tuning ability provided by >> YARN). >> >> Thanks, >> Mike >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Controlling-number-of-executors-on-Mesos-vs-YARN-tp20966.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >> >> >> >> >> -- >> Regards, >> Haripriya Ayyalasomayajula >> > > > > -- > Regards, > Haripriya Ayyalasomayajula > > > > > > -- > Regards, > Haripriya Ayyalasomayajula >