Github user mgummelt commented on a diff in the pull request:

    https://github.com/apache/spark/pull/10993#discussion_r51651630
  
    --- Diff: 
core/src/main/scala/org/apache/spark/scheduler/cluster/mesos/CoarseMesosSchedulerBackend.scala
 ---
    @@ -245,99 +239,182 @@ private[spark] class CoarseMesosSchedulerBackend(
        */
       override def resourceOffers(d: SchedulerDriver, offers: JList[Offer]) {
         stateLock.synchronized {
    -      val filters = Filters.newBuilder().setRefuseSeconds(5).build()
    -      for (offer <- offers.asScala) {
    +      logDebug(s"Received ${offers.size} resource offers.")
    +
    +      val (matchedOffers, unmatchedOffers) = offers.asScala.partition { 
offer =>
             val offerAttributes = toAttributeMap(offer.getAttributesList)
    -        val meetsConstraints = 
matchesAttributeRequirements(slaveOfferConstraints, offerAttributes)
    +        matchesAttributeRequirements(slaveOfferConstraints, 
offerAttributes)
    +      }
    +
    +      declineUnmatchedOffers(d, unmatchedOffers)
    +      handleMatchedOffers(d, matchedOffers)
    +    }
    +  }
    +
    +  private def declineUnmatchedOffers(d: SchedulerDriver, offers: 
Buffer[Offer]) {
    +    for (offer <- offers) {
    +      val id = offer.getId.getValue
    +      val offerAttributes = toAttributeMap(offer.getAttributesList)
    +      val mem = getResource(offer.getResourcesList, "mem")
    +      val cpus = getResource(offer.getResourcesList, "cpus")
    +      val filters = Filters.newBuilder()
    +        .setRefuseSeconds(rejectOfferDurationForUnmetConstraints).build()
    +
    +      logDebug(s"Declining offer: $id with attributes: $offerAttributes 
mem: $mem cpu: $cpus"
    +        + s" for $rejectOfferDurationForUnmetConstraints seconds")
    +
    +      d.declineOffer(offer.getId, filters)
    +    }
    +  }
    +
    +  private def handleMatchedOffers(d: SchedulerDriver, offers: 
Buffer[Offer]) {
    +    val tasks = getTasks(offers)
    +    for (offer <- offers) {
    +      val offerAttributes = toAttributeMap(offer.getAttributesList)
    +      val offerMem = getResource(offer.getResourcesList, "mem")
    +      val offerCpus = getResource(offer.getResourcesList, "cpus")
    +      val id = offer.getId.getValue
    +
    +      if (tasks.contains(offer.getId)) { // accept
    +        val filters = Filters.newBuilder().setRefuseSeconds(5).build()
    +        val offerTasks = tasks(offer.getId)
    +
    +        logDebug(s"Accepting offer: $id with attributes: $offerAttributes 
" +
    +          s"mem: $offerMem cpu: $offerCpus.  Launching ${offerTasks.size} 
Mesos tasks.")
    +
    +        for (task <- offerTasks) {
    +          val taskId = task.getTaskId
    +          val mem = getResource(task.getResourcesList, "mem")
    +          val cpus = getResource(task.getResourcesList, "cpus")
    +
    +          logDebug(s"Launching Mesos task: ${taskId.getValue} with mem: 
$mem cpu: $cpus.")
    +        }
    +
    +        d.launchTasks(
    +          Collections.singleton(offer.getId),
    +          offerTasks.asJava,
    +          filters)
    +      } else { // decline
    +        logDebug(s"Declining offer: $id with attributes: $offerAttributes 
" +
    +          s"mem: $offerMem cpu: $offerCpus")
    +
    +        d.declineOffer(offer.getId)
    +      }
    +    }
    +  }
    +
    +  private def getTasks(offers: Buffer[Offer]): mutable.Map[OfferID, 
List[MesosTaskInfo]] = {
    +    // offerID -> tasks
    +    val tasks = new HashMap[OfferID, 
List[MesosTaskInfo]].withDefaultValue(Nil)
    +
    +    // offerID -> resources
    +    val remainingResources = HashMap[String, 
JList[Resource]](offers.map(offer =>
    +      (offer.getId.getValue, offer.getResourcesList)): _*)
    +
    +    var launchTasks = true
    +
    +    // TODO(mgummelt): combine offers for a single slave
    +    // round-robin create executors on the available offers
    +    while (launchTasks) {
    +      launchTasks = false
    +
    +      for (offer <- offers) {
             val slaveId = offer.getSlaveId.getValue
    -        val mem = getResource(offer.getResourcesList, "mem")
    -        val cpus = getResource(offer.getResourcesList, "cpus").toInt
    -        val id = offer.getId.getValue
    -        if (meetsConstraints) {
    -          if (taskIdToSlaveId.size < executorLimit &&
    -              totalCoresAcquired < maxCores &&
    -              mem >= calculateTotalMemory(sc) &&
    -              cpus >= 1 &&
    -              failuresBySlaveId.getOrElse(slaveId, 0) < MAX_SLAVE_FAILURES 
&&
    -              !slaveIdsWithExecutors.contains(slaveId)) {
    -            // Launch an executor on the slave
    -            val cpusToUse = math.min(cpus, maxCores - totalCoresAcquired)
    -            totalCoresAcquired += cpusToUse
    -            val taskId = newMesosTaskId()
    -            taskIdToSlaveId.put(taskId, slaveId)
    -            slaveIdsWithExecutors += slaveId
    -            coresByTaskId(taskId) = cpusToUse
    -            // Gather cpu resources from the available resources and use 
them in the task.
    -            val (remainingResources, cpuResourcesToUse) =
    -              partitionResources(offer.getResourcesList, "cpus", cpusToUse)
    -            val (_, memResourcesToUse) =
    -              partitionResources(remainingResources.asJava, "mem", 
calculateTotalMemory(sc))
    -            val taskBuilder = MesosTaskInfo.newBuilder()
    -              
.setTaskId(TaskID.newBuilder().setValue(taskId.toString).build())
    -              .setSlaveId(offer.getSlaveId)
    -              .setCommand(createCommand(offer, cpusToUse + 
extraCoresPerSlave, taskId))
    -              .setName("Task " + taskId)
    -              .addAllResources(cpuResourcesToUse.asJava)
    -              .addAllResources(memResourcesToUse.asJava)
    -
    -            sc.conf.getOption("spark.mesos.executor.docker.image").foreach 
{ image =>
    -              MesosSchedulerBackendUtil
    -                .setupContainerBuilderDockerInfo(image, sc.conf, 
taskBuilder.getContainerBuilder())
    -            }
    -
    -            // Accept the offer and launch the task
    -            logDebug(s"Accepting offer: $id with attributes: 
$offerAttributes mem: $mem cpu: $cpus")
    -            slaveIdToHost(offer.getSlaveId.getValue) = offer.getHostname
    -            d.launchTasks(
    -              Collections.singleton(offer.getId),
    -              Collections.singleton(taskBuilder.build()), filters)
    -          } else {
    -            // Decline the offer
    -            logDebug(s"Declining offer: $id with attributes: 
$offerAttributes mem: $mem cpu: $cpus")
    -            d.declineOffer(offer.getId)
    +        val offerId = offer.getId.getValue
    +        val resources = remainingResources(offerId)
    +
    +        if (canLaunchTask(slaveId, resources)) {
    +          // Create a task
    +          launchTasks = true
    +          val taskId = newMesosTaskId()
    +          val offerCPUs = getResource(resources, "cpus").toInt
    +
    +          val taskCPUs = calculateTotalCPUs(offerCPUs)
    +          val taskMemory = calculateTotalMemory(sc)
    +
    +          slaves.getOrElseUpdate(slaveId, new 
Slave(offer.getHostname)).taskIDs.add(taskId)
    +
    +          val (afterCPUResources, cpuResourcesToUse) =
    +            partitionResources(resources, "cpus", taskCPUs)
    +          val (resourcesLeft, memResourcesToUse) =
    +            partitionResources(afterCPUResources.asJava, "mem", taskMemory)
    +
    +          val taskBuilder = MesosTaskInfo.newBuilder()
    +            
.setTaskId(TaskID.newBuilder().setValue(taskId.toString).build())
    +            .setSlaveId(offer.getSlaveId)
    +            .setCommand(createCommand(offer, taskCPUs + 
extraCoresPerExecutor, taskId))
    +            .setName("Task " + taskId)
    +            .addAllResources(cpuResourcesToUse.asJava)
    +            .addAllResources(memResourcesToUse.asJava)
    +
    +          sc.conf.getOption("spark.mesos.executor.docker.image").foreach { 
image =>
    +            MesosSchedulerBackendUtil
    +              .setupContainerBuilderDockerInfo(image, sc.conf, 
taskBuilder.getContainerBuilder)
               }
    -        } else {
    -          // This offer does not meet constraints. We don't need to see it 
again.
    -          // Decline the offer for a long period of time.
    -          logDebug(s"Declining offer: $id with attributes: 
$offerAttributes mem: $mem cpu: $cpus"
    -              + s" for $rejectOfferDurationForUnmetConstraints seconds")
    -          d.declineOffer(offer.getId, Filters.newBuilder()
    -            
.setRefuseSeconds(rejectOfferDurationForUnmetConstraints).build())
    +
    +          tasks(offer.getId) ::= taskBuilder.build()
    +          remainingResources(offerId) = resourcesLeft.asJava
    +          totalCoresAcquired += taskCPUs
    +          coresByTaskId(taskId) = taskCPUs
             }
           }
         }
    +    tasks
       }
     
    +  private def canLaunchTask(slaveId: String, resources: JList[Resource]): 
Boolean = {
    +    val offerMem = getResource(resources, "mem")
    +    val offerCPUs = getResource(resources, "cpus").toInt
    +    val cpus = calculateTotalCPUs(offerCPUs)
    +    val mem = calculateTotalMemory(sc)
    +
    +    cpus > 0 &&
    +      cpus <= offerCPUs &&
    +      cpus + totalCoresAcquired <= maxCores &&
    +      mem <= offerMem &&
    +      slaves.values.map(_.taskIDs.size).sum < executorLimit &&
    +      slaves.get(slaveId).map(_.taskFailures).getOrElse(0) < 
MAX_SLAVE_FAILURES
    +  }
    +
    +  private def calculateTotalCPUs(offerCPUs: Int): Int = {
    +    sc.conf.getInt("spark.executor.cores",
    +      math.min(offerCPUs, maxCores - totalCoresAcquired))
    +  }
     
       override def statusUpdate(d: SchedulerDriver, status: TaskStatus) {
    -    val taskId = status.getTaskId.getValue.toInt
    +    val taskId = status.getTaskId.getValue
         val state = status.getState
    +
         logInfo(s"Mesos task $taskId is now $state")
    +
         val slaveId: String = status.getSlaveId.getValue
    +    val slave = slaves(slaveId)
    +
         stateLock.synchronized {
           // If the shuffle service is enabled, have the driver register with 
each one of the
           // shuffle services. This allows the shuffle services to clean up 
state associated with
           // this application when the driver exits. There is currently not a 
great way to detect
           // this through Mesos, since the shuffle services are set up 
independently.
           if (TaskState.fromMesos(state).equals(TaskState.RUNNING) &&
    -          slaveIdToHost.contains(slaveId) &&
    -          shuffleServiceEnabled) {
    +          shuffleServiceEnabled &&
    +          !slave.shuffleRegistered) {
             assume(mesosExternalShuffleClient.isDefined,
               "External shuffle client was not instantiated even though 
shuffle service is enabled.")
             // TODO: Remove this and allow the MesosExternalShuffleService to 
detect
             // framework termination when new Mesos Framework HTTP API is 
available.
             val externalShufflePort = 
conf.getInt("spark.shuffle.service.port", 7337)
    -        val hostname = slaveIdToHost.remove(slaveId).get
    +
             logDebug(s"Connecting to shuffle service on slave $slaveId, " +
    -            s"host $hostname, port $externalShufflePort for app 
${conf.getAppId}")
    +            s"host ${slave.hostname}, port $externalShufflePort for app 
${conf.getAppId}")
    +
             mesosExternalShuffleClient.get
    -          .registerDriverWithShuffleService(hostname, externalShufflePort)
    +          .registerDriverWithShuffleService(slave.hostname, 
externalShufflePort)
    +        slave.shuffleRegistered = true
           }
     
           if (TaskState.isFinished(TaskState.fromMesos(state))) {
    -        val slaveId = taskIdToSlaveId.get(taskId)
    -        slaveIdsWithExecutors -= slaveId
    -        taskIdToSlaveId.remove(taskId)
    +        val slaveId = slaves.find(_._2.taskIDs.contains(taskId)).get._1
    --- End diff --
    
    I just realized that I already fetch the slave object above in this method, 
so I've removed this section. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to