beliefer commented on a change in pull request #27913: [SPARK-31002][CORE][DOC][FOLLOWUP] Add version information to the configuration of Core URL: https://github.com/apache/spark/pull/27913#discussion_r392579418
########## File path: core/src/main/scala/org/apache/spark/internal/config/package.scala ########## @@ -1187,48 +1247,56 @@ package object config { .internal() .doc("Initial threshold for the size of a collection before we start tracking its " + "memory usage.") + .version("1.1.1") .bytesConf(ByteUnit.BYTE) .createWithDefault(5 * 1024 * 1024) private[spark] val SHUFFLE_SPILL_BATCH_SIZE = ConfigBuilder("spark.shuffle.spill.batchSize") .internal() .doc("Size of object batches when reading/writing from serializers.") + .version("0.9.0") .longConf .createWithDefault(10000) private[spark] val SHUFFLE_SORT_BYPASS_MERGE_THRESHOLD = ConfigBuilder("spark.shuffle.sort.bypassMergeThreshold") .doc("In the sort-based shuffle manager, avoid merge-sorting data if there is no " + "map-side aggregation and there are at most this many reduce partitions") + .version("1.1.1") .intConf .createWithDefault(200) private[spark] val SHUFFLE_MANAGER = ConfigBuilder("spark.shuffle.manager") + .version("1.1.0") .stringConf .createWithDefault("sort") private[spark] val SHUFFLE_REDUCE_LOCALITY_ENABLE = ConfigBuilder("spark.shuffle.reduceLocality.enabled") .doc("Whether to compute locality preferences for reduce tasks") + .version("1.5.0") .booleanConf .createWithDefault(true) private[spark] val SHUFFLE_MAPOUTPUT_MIN_SIZE_FOR_BROADCAST = ConfigBuilder("spark.shuffle.mapOutput.minSizeForBroadcast") .doc("The size at which we use Broadcast to send the map output statuses to the executors.") + .version("2.0.0") .bytesConf(ByteUnit.BYTE) .createWithDefaultString("512k") private[spark] val SHUFFLE_MAPOUTPUT_DISPATCHER_NUM_THREADS = ConfigBuilder("spark.shuffle.mapOutput.dispatcher.numThreads") + .version("2.0.0") .intConf .createWithDefault(8) private[spark] val SHUFFLE_DETECT_CORRUPT = ConfigBuilder("spark.shuffle.detectCorrupt") .doc("Whether to detect any corruption in fetched blocks.") + .version("2.2.0") Review comment: SPARK-4105, commit ID: cf33a86285629abe72c1acf235b8bfa6057220a8#diff-eb30a71e0d04150b8e0b64929852e38b ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org