dongjoon-hyun commented on a change in pull request #23415: [SPARK-26445][CORE] Use ConfigEntry for hardcoded configs for driver/executor categories. URL: https://github.com/apache/spark/pull/23415#discussion_r244917210
########## File path: core/src/main/scala/org/apache/spark/ExecutorAllocationManager.scala ########## @@ -127,7 +127,7 @@ private[spark] class ExecutorAllocationManager( // allocation is only supported for YARN and the default number of cores per executor in YARN is // 1, but it might need to be attained differently for different cluster managers private val tasksPerExecutorForFullParallelism = - conf.getInt("spark.executor.cores", 1) / conf.getInt("spark.task.cpus", 1) + conf.get(EXECUTOR_CORES) / conf.getInt("spark.task.cpus", 1) Review comment: Hi, @kiszk . So, this seems to cause many conflicts with your PR because your PR chaning `spark.task.cpus`. Is it okay for us to proceed @ueshin 's PR first? ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org