beliefer commented on a change in pull request #27913: [SPARK-31002][CORE][DOC][FOLLOWUP] Add version information to the configuration of Core URL: https://github.com/apache/spark/pull/27913#discussion_r392578322
########## File path: core/src/main/scala/org/apache/spark/internal/config/package.scala ########## @@ -776,105 +776,128 @@ package object config { // This property sets the root namespace for metrics reporting private[spark] val METRICS_NAMESPACE = ConfigBuilder("spark.metrics.namespace") + .version("2.1.0") .stringConf .createOptional private[spark] val METRICS_CONF = ConfigBuilder("spark.metrics.conf") + .version("0.8.0") .stringConf .createOptional private[spark] val METRICS_EXECUTORMETRICS_SOURCE_ENABLED = ConfigBuilder("spark.metrics.executorMetricsSource.enabled") .doc("Whether to register the ExecutorMetrics source with the metrics system.") + .version("3.0.0") .booleanConf .createWithDefault(true) private[spark] val METRICS_STATIC_SOURCES_ENABLED = ConfigBuilder("spark.metrics.staticSources.enabled") .doc("Whether to register static sources with the metrics system.") + .version("3.0.0") .booleanConf .createWithDefault(true) private[spark] val PYSPARK_DRIVER_PYTHON = ConfigBuilder("spark.pyspark.driver.python") + .version("2.1.0") .stringConf .createOptional private[spark] val PYSPARK_PYTHON = ConfigBuilder("spark.pyspark.python") + .version("2.1.0") .stringConf .createOptional // To limit how many applications are shown in the History Server summary ui private[spark] val HISTORY_UI_MAX_APPS = - ConfigBuilder("spark.history.ui.maxApplications").intConf.createWithDefault(Integer.MAX_VALUE) + ConfigBuilder("spark.history.ui.maxApplications") + .version("2.0.1") + .intConf + .createWithDefault(Integer.MAX_VALUE) private[spark] val IO_ENCRYPTION_ENABLED = ConfigBuilder("spark.io.encryption.enabled") + .version("2.1.0") .booleanConf .createWithDefault(false) private[spark] val IO_ENCRYPTION_KEYGEN_ALGORITHM = ConfigBuilder("spark.io.encryption.keygen.algorithm") + .version("2.1.0") .stringConf .createWithDefault("HmacSHA1") private[spark] val IO_ENCRYPTION_KEY_SIZE_BITS = ConfigBuilder("spark.io.encryption.keySizeBits") + .version("2.1.0") .intConf .checkValues(Set(128, 192, 256)) .createWithDefault(128) private[spark] val IO_CRYPTO_CIPHER_TRANSFORMATION = ConfigBuilder("spark.io.crypto.cipher.transformation") .internal() + .version("2.1.0") .stringConf .createWithDefaultString("AES/CTR/NoPadding") private[spark] val DRIVER_HOST_ADDRESS = ConfigBuilder("spark.driver.host") .doc("Address of driver endpoints.") + .version("0.7.0") .stringConf .createWithDefault(Utils.localCanonicalHostName()) private[spark] val DRIVER_PORT = ConfigBuilder("spark.driver.port") .doc("Port of driver endpoints.") + .version("0.7.0") .intConf .createWithDefault(0) private[spark] val DRIVER_SUPERVISE = ConfigBuilder("spark.driver.supervise") .doc("If true, restarts the driver automatically if it fails with a non-zero exit status. " + "Only has effect in Spark standalone mode or Mesos cluster deploy mode.") + .version("1.3.0") .booleanConf .createWithDefault(false) private[spark] val DRIVER_BIND_ADDRESS = ConfigBuilder("spark.driver.bindAddress") .doc("Address where to bind network listen sockets on the driver.") + .version("2.1.0") Review comment: SPARK-4563, commit ID: 2cd1bfa4f0c6625b0ab1dbeba2b9586b9a6a9f42#diff-6bdad48cfc34314e89599655442ff210 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org