I have a spark cluster running in client mode with driver outside the spark cluster. I want to scale the cluster after an application is submitted. In order to do this, I'm creating new workers and they are getting registered with master but issue I'm seeing is; running application does not use the newly added worker. Hence cannot add more resources to existing running application.
Is there any other way or config to deal with this use-case ? How to make running application to ask for executors from newly issued worker node ? -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Scaling-spark-cluster-for-a-running-application-tp23951.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org