Github user cloud-fan commented on a diff in the pull request: https://github.com/apache/spark/pull/18555#discussion_r126274198 --- Diff: core/src/main/scala/org/apache/spark/internal/config/package.scala --- @@ -88,50 +137,76 @@ package object config { .createOptional private[spark] val PY_FILES = ConfigBuilder("spark.submit.pyFiles") + .doc("Comma-separated list of .zip, .egg, or .py files to place on " + + "the PYTHONPATH for Python apps.") .internal() .stringConf .toSequence .createWithDefault(Nil) private[spark] val MAX_TASK_FAILURES = ConfigBuilder("spark.task.maxFailures") + .doc("Number of failures of any particular task before giving up on the job. " + + "The total number of failures spread across different tasks " + + "will not cause the job to fail; " + + "a particular task has to fail this number of attempts. " + + "Should be greater than or equal to 1. Number of allowed retries = this value - 1.") .intConf + .checkValue(_ > 0, "The retry times of task should be greater than or equal to 1") .createWithDefault(4) // Blacklist confs private[spark] val BLACKLIST_ENABLED = ConfigBuilder("spark.blacklist.enabled") + .doc("If set to 'true', prevent Spark from scheduling tasks on executors " + + "that have been blacklisted due to too many task failures. " + + "The blacklisting algorithm can be further controlled by " + + "the other 'spark.blacklist' configuration options. ") .booleanConf .createOptional private[spark] val MAX_TASK_ATTEMPTS_PER_EXECUTOR = ConfigBuilder("spark.blacklist.task.maxTaskAttemptsPerExecutor") + .doc("For a given task, how many times it can be " + + "retried on one executor before the executor is blacklisted for that task.") .intConf + .checkValue(_ > 0, "The value should be greater than or equal to 1") --- End diff -- `should be greater than 0`
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org