Kimahriman commented on code in PR #47393: URL: https://github.com/apache/spark/pull/47393#discussion_r1681869279
########## sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala: ########## @@ -2129,6 +2129,13 @@ object SQLConf { .intConf .createWithDefault(100) + val MIN_VERSIONS_TO_DELETE = buildConf("spark.sql.streaming.minVersionsToDelete") + .internal() + .doc("The minimum number of stale versions to delete when maintenance is invoked.") + .version("2.1.1") + .intConf + .createWithDefault(30) Review Comment: Yeah I'm not really sure why the default minBatchesToRetain is so high, we set it to 3 in our jobs hah. I would likely set this new config to 1 if that's not the default, I'm just not sure how many other rocksdb state users would be impacted by effectively maintaining more state store versions by default. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org