> by specifying a larger heap size than default on each worker node. I don't follow. Which heap? Are you specifying a large heap size on the executors? If so, do you mean you somehow launch the shuffle service when you launch executors? Or something else?
On Wed, Feb 8, 2017 at 5:50 PM, Sun Rui <sunrise_...@163.com> wrote: > Michael, > No. We directly launch the external shuffle service by specifying a larger > heap size than default on each worker node. It is observed that the > processes are quite stable. > > On Feb 9, 2017, at 05:21, Michael Gummelt <mgumm...@mesosphere.io> wrote: > > Sun, are you using marathon to run the shuffle service? > > On Tue, Feb 7, 2017 at 7:36 PM, Sun Rui <sunrise_...@163.com> wrote: > >> Yi Jan, >> >> We have been using Spark on Mesos with dynamic allocation enabled, which >> works and improves the overall cluster utilization. >> >> In terms of job, do you mean jobs inside a Spark application or jobs >> among different applications? Maybe you can read >> http://spark.apache.org/docs/latest/job-scheduling.html for help. >> >> On Jan 31, 2017, at 03:34, Michael Gummelt <mgumm...@mesosphere.io> >> wrote: >> >> >> >> On Mon, Jan 30, 2017 at 9:47 AM, Ji Yan <ji...@drive.ai> wrote: >> >>> Tasks begin scheduling as soon as the first executor comes up >>> >>> >>> Thanks all for the clarification. Is this the default behavior of Spark >>> on Mesos today? I think this is what we are looking for because sometimes a >>> job can take up lots of resources and later jobs could not get all the >>> resources that it asks for. If a Spark job starts with only a subset of >>> resources that it asks for, does it know to expand its resources later when >>> more resources become available? >>> >> >> Yes. >> >> >>> >>> Launch each executor with at least 1GB RAM, but if mesos offers 2GB at >>>> some moment, then launch an executor with 2GB RAM >>> >>> >>> This is less useful in our use case. But I am also quite interested in >>> cases in which this could be helpful. I think this will also help with >>> overall resource utilization on the cluster if when another job starts up >>> that has a hard requirement on resources, the extra resources to the first >>> job can be flexibly re-allocated to the second job. >>> >>> On Sat, Jan 28, 2017 at 2:32 PM, Michael Gummelt <mgumm...@mesosphere.io >>> > wrote: >>> >>>> We've talked about that, but it hasn't become a priority because we >>>> haven't had a driving use case. If anyone has a good argument for >>>> "variable" resource allocation like this, please let me know. >>>> >>>> On Sat, Jan 28, 2017 at 9:17 AM, Shuai Lin <linshuai2...@gmail.com> w >>>> rote: >>>> >>>>> An alternative behavior is to launch the job with the best resource >>>>>> offer Mesos is able to give >>>>> >>>>> >>>>> Michael has just made an excellent explanation about dynamic >>>>> allocation support in mesos. But IIUC, what you want to achieve is >>>>> something like (using RAM as an example) : "Launch each executor with at >>>>> least 1GB RAM, but if mesos offers 2GB at some moment, then launch an >>>>> executor with 2GB RAM". >>>>> >>>>> I wonder what's benefit of that? To reduce the "resource >>>>> fragmentation"? >>>>> >>>>> Anyway, that is not supported at this moment. In all the supported >>>>> cluster managers of spark (mesos, yarn, standalone, and the up-to-coming >>>>> spark on kubernetes), you have to specify the cores and memory of each >>>>> executor. >>>>> >>>>> It may not be supported in the future, because only mesos has the >>>>> concepts of offers because of its two-level scheduling model. >>>>> >>>>> >>>>> On Sat, Jan 28, 2017 at 1:35 AM, Ji Yan <ji...@drive.ai> wrote: >>>>> >>>>>> Dear Spark Users, >>>>>> >>>>>> Currently is there a way to dynamically allocate resources to Spark >>>>>> on Mesos? Within Spark we can specify the CPU cores, memory before >>>>>> running >>>>>> job. The way I understand is that the Spark job will not run if the >>>>>> CPU/Mem >>>>>> requirement is not met. This may lead to decrease in overall utilization >>>>>> of >>>>>> the cluster. An alternative behavior is to launch the job with the best >>>>>> resource offer Mesos is able to give. Is this possible with the current >>>>>> implementation? >>>>>> >>>>>> Thanks >>>>>> Ji >>>>>> >>>>>> The information in this email is confidential and may be legally >>>>>> privileged. It is intended solely for the addressee. Access to this email >>>>>> by anyone else is unauthorized. If you are not the intended recipient, >>>>>> any >>>>>> disclosure, copying, distribution or any action taken or omitted to be >>>>>> taken in reliance on it, is prohibited and may be unlawful. >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Michael Gummelt >>>> Software Engineer >>>> Mesosphere >>>> >>> >>> >>> The information in this email is confidential and may be legally >>> privileged. It is intended solely for the addressee. Access to this email >>> by anyone else is unauthorized. If you are not the intended recipient, any >>> disclosure, copying, distribution or any action taken or omitted to be >>> taken in reliance on it, is prohibited and may be unlawful. >>> >> >> >> >> -- >> Michael Gummelt >> Software Engineer >> Mesosphere >> >> >> > > > -- > Michael Gummelt > Software Engineer > Mesosphere > > > -- Michael Gummelt Software Engineer Mesosphere