Github user LantaoJin commented on a diff in the pull request: https://github.com/apache/spark/pull/20532#discussion_r166860784 --- Diff: core/src/main/scala/org/apache/spark/internal/config/package.scala --- @@ -53,10 +53,21 @@ package object config { .booleanConf .createWithDefault(false) - private[spark] val EVENT_LOG_BLOCK_UPDATES = - ConfigBuilder("spark.eventLog.logBlockUpdates.enabled") - .booleanConf - .createWithDefault(false) + private[spark] val EVENT_LOG_BLOCK_UPDATES_FRACTION = + ConfigBuilder("spark.eventLog.logBlockUpdates.fraction") + .doc("Expected number of times each blockUpdated event is chosen to log, " + + "fraction must be [0, 1]. 0 by default, means disabled") + .doubleConf + .checkValue(_ >= 0, "The fraction must not be negative") --- End diff -- Yes. I know a custom SparkListener can do the dump. But actually we can dump to eventlog file. So the conclusion is that sample is unacceptable? Is adding a configuration like "spark.eventLog.logVerboseEvent.enabled" to enable it enough?
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org