Github user vanzin commented on a diff in the pull request: https://github.com/apache/spark/pull/14943#discussion_r77722666 --- Diff: yarn/src/main/scala/org/apache/spark/deploy/yarn/ExecutorRunnable.scala --- @@ -59,43 +58,46 @@ private[yarn] class ExecutorRunnable( var rpc: YarnRPC = YarnRPC.create(conf) var nmClient: NMClient = _ - val yarnConf: YarnConfiguration = new YarnConfiguration(conf) - lazy val env = prepareEnvironment(container) def run(): Unit = { - logInfo("Starting Executor Container") + logDebug("Starting Executor Container") nmClient = NMClient.createNMClient() - nmClient.init(yarnConf) + nmClient.init(conf) nmClient.start() startContainer() } - def startContainer(): java.util.Map[String, ByteBuffer] = { - logInfo("Setting up ContainerLaunchContext") + def launchContextDebugInfo(): String = { + val commands = prepareCommand() --- End diff -- I thought about adding a boolean "has the context been logged?" to control when to log. But then it would contain information about the first executor only, and this allows be to put placeholders for executor id and hostname, for example. So the log message looks a little better even if the code is not optimal. As for the command / env, the command is mostly the same, aside from the exec id and hostname; so every launcher would still need to do some customization.
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. --- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org