Github user vanzin commented on a diff in the pull request:

    https://github.com/apache/spark/pull/14079#discussion_r70005272
  
    --- Diff: 
core/src/main/scala/org/apache/spark/scheduler/BlacklistTracker.scala ---
    @@ -0,0 +1,320 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *    http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +
    +package org.apache.spark.scheduler
    +
    +import java.util.concurrent.atomic.AtomicReference
    +
    +import scala.collection.mutable.{HashMap, HashSet}
    +
    +import org.apache.spark.SparkConf
    +import org.apache.spark.internal.Logging
    +import org.apache.spark.util.Clock
    +import org.apache.spark.util.SystemClock
    +import org.apache.spark.util.Utils
    +
    +/**
    + * BlacklistTracker is designed to track problematic executors and nodes.  
It supports blacklisting
    + * specific (executor, task) pairs within a stage, blacklisting entire 
executors and nodes for a
    + * stage, and blacklisting executors and nodes across an entire 
application (with a periodic
    + * expiry).
    + *
    + * The tracker needs to deal with a variety of workloads, eg.: bad user 
code, which may lead to many
    + * task failures, but that should not count against individual executors; 
many small stages, which
    + * may prevent a bad executor for having many failures within one stage, 
but still many failures
    + * over the entire application; "flaky" executors, that don't fail every 
task, but are still
    + * faulty; etc.
    + *
    + * THREADING: As with most helpers of TaskSchedulerImpl, this is not 
thread-safe.  Though it is
    +  * called by multiple threads, callers must already have a lock on the 
TaskSchedulerImpl.  The
    +  * one exception is [[nodeBlacklist()]], which can be called without 
holding a lock.
    + */
    +private[scheduler] class BlacklistTracker (
    +    conf: SparkConf,
    +    clock: Clock = new SystemClock()) extends Logging {
    +
    +  private val MAX_TASK_FAILURES_PER_NODE =
    +    conf.getInt("spark.blacklist.maxTaskFailuresPerNode", 2)
    +  private val MAX_FAILURES_PER_EXEC =
    +    conf.getInt("spark.blacklist.maxFailedTasksPerExecutor", 2)
    +  private val MAX_FAILURES_PER_EXEC_STAGE =
    +    conf.getInt("spark.blacklist.maxFailedTasksPerExecutorStage", 2)
    +  private val MAX_FAILED_EXEC_PER_NODE =
    +    conf.getInt("spark.blacklist.maxFailedExecutorsPerNode", 2)
    +  private val MAX_FAILED_EXEC_PER_NODE_STAGE =
    +    conf.getInt("spark.blacklist.maxFailedExecutorsPerNodeStage", 2)
    +  val EXECUTOR_RECOVERY_MILLIS = 
BlacklistTracker.getBlacklistExpiryTime(conf)
    +
    +  // a count of failed tasks for each executor.  Only counts failures 
after tasksets complete
    +  // successfully
    +  private val executorIdToFailureCount: HashMap[String, Int] = HashMap()
    +  // failures for each executor by stage.  Only tracked while the stage is 
running.
    +  val stageIdToExecToFailures: HashMap[Int, HashMap[String, 
FailureStatus]] =
    +    new HashMap()
    +  val stageIdToNodeBlacklistedTasks: HashMap[Int, HashMap[String, 
HashSet[Int]]] =
    +    new HashMap()
    +  val stageIdToBlacklistedNodes: HashMap[Int, HashSet[String]] = new 
HashMap()
    +  private val executorIdToBlacklistExpiryTime: HashMap[String, Long] = new 
HashMap()
    +  private val nodeIdToBlacklistExpiryTime: HashMap[String, Long] = new 
HashMap()
    +  private val _nodeBlacklist: AtomicReference[Set[String]] = new 
AtomicReference(Set())
    +  private var nextExpiryTime: Long = Long.MaxValue
    +
    +  def start(): Unit = {}
    +
    +  def stop(): Unit = {}
    +
    +  def expireExecutorsInBlacklist(): Unit = {
    +    val now = clock.getTimeMillis()
    +    // quickly check if we've got anything to expire from blacklist -- if 
not, avoid doing any work
    +    if (now > nextExpiryTime) {
    +      val execsToClear = executorIdToBlacklistExpiryTime.filter(_._2 < 
now).keys
    +      if (execsToClear.nonEmpty) {
    +        logInfo(s"Removing executors $execsToClear from blacklist during 
periodic recovery")
    +        execsToClear.foreach { exec => 
executorIdToBlacklistExpiryTime.remove(exec) }
    +      }
    +      if (executorIdToBlacklistExpiryTime.nonEmpty) {
    +        nextExpiryTime = executorIdToBlacklistExpiryTime.map{_._2}.min
    +      } else {
    +        nextExpiryTime = Long.MaxValue
    +      }
    +      val nodesToClear = nodeIdToBlacklistExpiryTime.filter(_._2 < 
now).keys
    +      if (nodesToClear.nonEmpty) {
    +        logInfo(s"Removing nodes $nodesToClear from blacklist during 
periodic recovery")
    +        nodesToClear.foreach { node => 
nodeIdToBlacklistExpiryTime.remove(node) }
    +        // make a copy of the blacklisted nodes so nodeBlacklist() is 
threadsafe
    +        _nodeBlacklist.set(nodeIdToBlacklistExpiryTime.keySet.toSet)
    +      }
    +    }
    + }
    +
    +  def taskSetSucceeded(stageId: Int, scheduler: TaskSchedulerImpl): Unit = 
{
    +    // if any tasks failed, we count them towards the overall failure 
count for the executor at
    +    // this point.  Also clean out all data about the stage to avoid 
increasing memory use.
    +    stageIdToExecToFailures.remove(stageId).map { failuresForStage =>
    +      failuresForStage.foreach { case (exec, newFailures) =>
    +        val prevFailures = executorIdToFailureCount.getOrElse(exec, 0)
    +        val newTotal = prevFailures + newFailures.totalFailures
    +
    +        if (newTotal >= MAX_FAILURES_PER_EXEC) {
    +          logInfo(s"Blacklisting executor $exec because it has $newTotal" +
    +            s" task failures in successful task sets")
    +          val now = clock.getTimeMillis()
    +          val expiryTime = now + EXECUTOR_RECOVERY_MILLIS
    +          executorIdToBlacklistExpiryTime.put(exec, expiryTime)
    +          executorIdToFailureCount.remove(exec)
    +          if (expiryTime < nextExpiryTime) {
    +            nextExpiryTime = expiryTime
    +          }
    +
    +          val node = scheduler.getHostForExecutor(exec)
    +          val execs = 
scheduler.getExecutorsAliveOnHost(node).getOrElse(Set())
    +          val blacklistedExecs = 
execs.filter(executorIdToBlacklistExpiryTime.contains(_))
    +          if (blacklistedExecs.size >= MAX_FAILED_EXEC_PER_NODE) {
    +            logInfo(s"Blacklisting node $node because it has 
${blacklistedExecs.size} executors " +
    +              s"blacklisted: ${blacklistedExecs}")
    +            nodeIdToBlacklistExpiryTime.put(node, expiryTime)
    +            // make a copy of the blacklisted nodes so nodeBlacklist() is 
threadsafe
    +            _nodeBlacklist.set(nodeIdToBlacklistExpiryTime.keySet.toSet)
    +          }
    +        } else {
    +          executorIdToFailureCount.put(exec, newTotal)
    +        }
    +      }
    +    }
    +    // when we blacklist a node within a stage, we don't directly promote 
that node to being
    +    // blacklisted for the app.  Instead, we use the mechanism above to 
decide whether or not to
    +    // blacklist any executors for the app, and when doing so we'll check 
whether or not to also
    +    // blacklist the node.  That is why we just remove this entry without 
doing any promotion to
    +    // the full app blacklist.
    +    stageIdToBlacklistedNodes.remove(stageId)
    +  }
    +
    +  def taskSetFailed(stageId: Int): Unit = {
    +    // just throw away all the info for the failures in this taskSet -- 
assume the executors were
    +    // fine, the failures were just b/c the taskSet itself was bad (eg., 
bad user code)
    +    stageIdToExecToFailures.remove(stageId)
    +    stageIdToBlacklistedNodes.remove(stageId)
    +  }
    +
    +  /**
    +   * Return true if this executor is blacklisted for the given stage.  
Completely ignores whether
    +   * the executor is blacklisted overall (or anything to do with the node 
the executor is on).
    +   */
    +  def isExecutorBlacklistedForStage(
    +      stageId: Int,
    +      executorId: String): Boolean = {
    +    stageIdToExecToFailures.get(stageId).flatMap(_.get(executorId))
    +      .map(_.totalFailures >= MAX_FAILURES_PER_EXEC_STAGE).getOrElse(false)
    +  }
    +
    +  def isExecutorBlacklisted(executorId: String): Boolean = {
    +    executorIdToBlacklistExpiryTime.contains(executorId)
    +  }
    +
    +  def isNodeBlacklistedForStage(node: String, stageId: Int): Boolean = {
    +    
stageIdToBlacklistedNodes.get(stageId).map(_.contains(node)).getOrElse(false)
    +  }
    +
    +  def nodeBlacklist(): Set[String] = {
    +    _nodeBlacklist.get()
    +  }
    +
    +  def isNodeBlacklisted(node: String): Boolean = {
    +    nodeIdToBlacklistExpiryTime.contains(node)
    +  }
    +
    +  def taskSucceeded(
    +      stageId: Int,
    +      indexInTaskSet: Int,
    +      info: TaskInfo,
    +      scheduler: TaskSchedulerImpl): Unit = {
    +    // no-op intentionally, included just for symmetry.  success to 
failure ratio is irrelevant, we
    +    // just blacklist based on failures.  Furthermore, one success does 
not previous
    +    // failures, since the bad node / executor may not fail *every* time
    +  }
    +
    +  def taskFailed(
    +      stageId: Int,
    +      indexInTaskSet: Int,
    +      info: TaskInfo,
    +      scheduler: TaskSchedulerImpl): Unit = {
    +    val stageFailures = stageIdToExecToFailures.getOrElseUpdate(stageId, 
new HashMap())
    +    val failureStatus = stageFailures.getOrElseUpdate(info.executorId, new 
FailureStatus())
    +    failureStatus.totalFailures += 1
    +    failureStatus.failuresByTask += indexInTaskSet
    +
    +    // check if this task has also failed on other executors on the same 
host, and if so, blacklist
    +    // this task from the host
    +    val failuresOnHost = (for {
    +      exec <- 
scheduler.getExecutorsAliveOnHost(info.host).getOrElse(Set()).toSeq
    +      failures <- stageFailures.get(exec)
    +    } yield {
    +      if (failures.failuresByTask.contains(indexInTaskSet)) 1 else 0
    +    }).sum
    +    logInfo(s"total failures on host ${info.host} = $failuresOnHost")
    +    if (failuresOnHost > MAX_TASK_FAILURES_PER_NODE) {
    +      stageIdToNodeBlacklistedTasks.getOrElseUpdate(stageId, new HashMap())
    +        .getOrElseUpdate(info.host, new HashSet()) += indexInTaskSet
    +    }
    +
    +
    +    if (failureStatus.totalFailures >= MAX_FAILURES_PER_EXEC_STAGE) {
    +      // this executor has been pushed into the blacklist for this stage.  
Lets check if it pushes
    +      // the whole node into the blacklist
    +      val blacklistedExecutors =
    +        stageFailures.filter{_._2.totalFailures >= 
MAX_FAILURES_PER_EXEC_STAGE}
    +      if (blacklistedExecutors.size >= MAX_FAILED_EXEC_PER_NODE_STAGE) {
    +        logInfo(s"Blacklisting ${info.host} for stage $stageId")
    +        stageIdToBlacklistedNodes.getOrElseUpdate(stageId, new HashSet()) 
+= info.host
    +      }
    +    }
    +  }
    +
    +  /**
    +   * Return true if this executor is blacklisted for the given task.  This 
does *not*
    +   * need to return true if the executor is blacklisted for the entire 
stage, or blacklisted
    +   * altogether.
    +   */
    +  def isExecutorBlacklisted(
    +      executorId: String,
    +      stageId: Int,
    +      indexInTaskSet: Int): Boolean = {
    +    // intentionally avoiding .getOrElse(..., new HashMap()) to avoid lots 
of object
    +    // creation, since this method gets called a *lot*
    +    stageIdToExecToFailures.get(stageId) match {
    +      case Some(stageFailures) =>
    +        stageFailures.get(executorId) match {
    +          case Some(failures) =>
    +            failures.failuresByTask.contains(indexInTaskSet)
    +          case None =>
    +            false
    +        }
    +      case None =>
    +        false
    +    }
    +  }
    +
    +  def isNodeBlacklisted(
    +      node: String,
    +      stageId: Int,
    +      indexInTaskSet: Int): Boolean = {
    +    stageIdToNodeBlacklistedTasks.get(stageId).flatMap { nodeToFailures =>
    +      nodeToFailures.get(node).map{_.contains(indexInTaskSet)}
    +    }.getOrElse(false)
    +  }
    +
    +  def removeExecutor(executorId: String): Unit = {
    +    executorIdToBlacklistExpiryTime -= executorId
    +    executorIdToFailureCount -= executorId
    +    stageIdToExecToFailures.values.foreach { execFailureOneStage =>
    +      execFailureOneStage -= executorId
    +    }
    +  }
    +}
    +
    +
    +private[scheduler] object BlacklistTracker extends Logging {
    +  val LEGACY_TIMEOUT_CONF = "spark.scheduler.executorTaskBlacklistTime"
    +  val EXPIRY_TIMEOUT_CONF = "spark.scheduler.blacklist.recoverPeriod"
    +  val ENABLED_CONF = "spark.scheduler.blacklist.enabled"
    +
    +  /**
    +   * Return true if the blacklist is enabled, based on the following order 
of preferences:
    +   * 1. Is it specifically enabled or disabled?
    +   * 2. Is it enabled via the legacy timeout conf?
    +   * 3. Use the default for the spark-master:
    +   *   - off for local mode
    +   *   - on for distributed modes (including local-cluster)
    +   */
    +  def isBlacklistEnabled(conf: SparkConf): Boolean = {
    +    val isEnabled = conf.get(ENABLED_CONF, null)
    +    if (isEnabled == null) {
    +      // if they've got a non-zero setting for the legacy conf, always 
enable the blacklist,
    +      // otherwise, use the default based on the cluster-mode (off for 
local-mode, on otherwise).
    +      val legacyTimeout = conf.getLong(LEGACY_TIMEOUT_CONF, 0L)
    +      if (legacyTimeout > 0) {
    +        // mostly this is necessary just for tests, since real users that 
want the blacklist will
    +        // get it anyway by default
    +        logWarning(s"Turning on blacklisting due to legacy configuration: 
$LEGACY_TIMEOUT_CONF > 0")
    +        true
    +      } else {
    +        // local-cluster is *not* considered local for these purposes, we 
still want the blacklist
    +        // enabled by default
    +        !Utils.isLocalMaster(conf)
    +      }
    +    } else {
    +      // always take whatever value is explicitly set by the user
    +      isEnabled.toBoolean
    +    }
    +  }
    +
    +  def getBlacklistExpiryTime(conf: SparkConf): Long = {
    +    conf.getTimeAsMs(BlacklistTracker.EXPIRY_TIMEOUT_CONF,
    +      conf.get(BlacklistTracker.LEGACY_TIMEOUT_CONF, (60 * 60 * 
1000).toString))
    --- End diff --
    
    `"1h"` instead of doing math.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to