Sun, are you using marathon to run the shuffle service? On Tue, Feb 7, 2017 at 7:36 PM, Sun Rui <sunrise_...@163.com> wrote:
> Yi Jan, > > We have been using Spark on Mesos with dynamic allocation enabled, which > works and improves the overall cluster utilization. > > In terms of job, do you mean jobs inside a Spark application or jobs among > different applications? Maybe you can read http://spark.apache.org/ > docs/latest/job-scheduling.html for help. > > On Jan 31, 2017, at 03:34, Michael Gummelt <mgumm...@mesosphere.io> wrote: > > > > On Mon, Jan 30, 2017 at 9:47 AM, Ji Yan <ji...@drive.ai> wrote: > >> Tasks begin scheduling as soon as the first executor comes up >> >> >> Thanks all for the clarification. Is this the default behavior of Spark >> on Mesos today? I think this is what we are looking for because sometimes a >> job can take up lots of resources and later jobs could not get all the >> resources that it asks for. If a Spark job starts with only a subset of >> resources that it asks for, does it know to expand its resources later when >> more resources become available? >> > > Yes. > > >> >> Launch each executor with at least 1GB RAM, but if mesos offers 2GB at >>> some moment, then launch an executor with 2GB RAM >> >> >> This is less useful in our use case. But I am also quite interested in >> cases in which this could be helpful. I think this will also help with >> overall resource utilization on the cluster if when another job starts up >> that has a hard requirement on resources, the extra resources to the first >> job can be flexibly re-allocated to the second job. >> >> On Sat, Jan 28, 2017 at 2:32 PM, Michael Gummelt <mgumm...@mesosphere.io> >> wrote: >> >>> We've talked about that, but it hasn't become a priority because we >>> haven't had a driving use case. If anyone has a good argument for >>> "variable" resource allocation like this, please let me know. >>> >>> On Sat, Jan 28, 2017 at 9:17 AM, Shuai Lin <linshuai2...@gmail.com> w >>> rote: >>> >>>> An alternative behavior is to launch the job with the best resource >>>>> offer Mesos is able to give >>>> >>>> >>>> Michael has just made an excellent explanation about dynamic allocation >>>> support in mesos. But IIUC, what you want to achieve is something like >>>> (using RAM as an example) : "Launch each executor with at least 1GB RAM, >>>> but if mesos offers 2GB at some moment, then launch an executor with 2GB >>>> RAM". >>>> >>>> I wonder what's benefit of that? To reduce the "resource fragmentation"? >>>> >>>> Anyway, that is not supported at this moment. In all the supported >>>> cluster managers of spark (mesos, yarn, standalone, and the up-to-coming >>>> spark on kubernetes), you have to specify the cores and memory of each >>>> executor. >>>> >>>> It may not be supported in the future, because only mesos has the >>>> concepts of offers because of its two-level scheduling model. >>>> >>>> >>>> On Sat, Jan 28, 2017 at 1:35 AM, Ji Yan <ji...@drive.ai> wrote: >>>> >>>>> Dear Spark Users, >>>>> >>>>> Currently is there a way to dynamically allocate resources to Spark on >>>>> Mesos? Within Spark we can specify the CPU cores, memory before running >>>>> job. The way I understand is that the Spark job will not run if the >>>>> CPU/Mem >>>>> requirement is not met. This may lead to decrease in overall utilization >>>>> of >>>>> the cluster. An alternative behavior is to launch the job with the best >>>>> resource offer Mesos is able to give. Is this possible with the current >>>>> implementation? >>>>> >>>>> Thanks >>>>> Ji >>>>> >>>>> The information in this email is confidential and may be legally >>>>> privileged. It is intended solely for the addressee. Access to this email >>>>> by anyone else is unauthorized. If you are not the intended recipient, any >>>>> disclosure, copying, distribution or any action taken or omitted to be >>>>> taken in reliance on it, is prohibited and may be unlawful. >>>>> >>>> >>>> >>> >>> >>> -- >>> Michael Gummelt >>> Software Engineer >>> Mesosphere >>> >> >> >> The information in this email is confidential and may be legally >> privileged. It is intended solely for the addressee. Access to this email >> by anyone else is unauthorized. If you are not the intended recipient, any >> disclosure, copying, distribution or any action taken or omitted to be >> taken in reliance on it, is prohibited and may be unlawful. >> > > > > -- > Michael Gummelt > Software Engineer > Mesosphere > > > -- Michael Gummelt Software Engineer Mesosphere