start-all start the master and anything else in slaves file start-master.sh starts the master only.
I use start-slaves.sh for my purpose with added nodes to slaves file. When you run start-slave.sh <MASTER_IP_ADD> you are creating another worker process on the master host. You can check the status on Spark GUI on <HOST>:8080. Depending the ratio of Memory/core for worker process the additional worker may or may not be used. Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com On 28 March 2016 at 22:58, Sung Hwan Chung <coded...@cs.stanford.edu> wrote: > It seems that the conf/slaves file is only for consumption by the > following scripts: > > sbin/start-slaves.sh > sbin/stop-slaves.sh > sbin/start-all.sh > sbin/stop-all.sh > > I.e., conf/slaves file doesn't affect a running cluster. > > Is this true? > > > On Mon, Mar 28, 2016 at 9:31 PM, Sung Hwan Chung <coded...@cs.stanford.edu > > wrote: > >> No I didn't add it to the conf/slaves file. >> >> What I want to do is leverage auto-scale from AWS, without needing to >> stop all the slaves (e.g. if a lot of slaves are idle, terminate those). >> >> Also, the book-keeping is easier if I don't have to deal with some >> centralized list of slave list that needs to be modified every time a node >> is added/removed. >> >> >> On Mon, Mar 28, 2016 at 9:20 PM, Mich Talebzadeh < >> mich.talebza...@gmail.com> wrote: >> >>> Have you added the slave host name to $SPARK_HOME/conf? >>> >>> Then you can use start-slaves.sh or stop-slaves.sh for all instances >>> >>> The assumption is that slave boxes have $SPARK_HOME installed in the >>> same directory as $SPARK_HOME is installed in the master. >>> >>> HTH >>> >>> >>> Dr Mich Talebzadeh >>> >>> >>> >>> LinkedIn * >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>> >>> >>> >>> http://talebzadehmich.wordpress.com >>> >>> >>> >>> On 28 March 2016 at 22:06, Sung Hwan Chung <coded...@cs.stanford.edu> >>> wrote: >>> >>>> Hello, >>>> >>>> I found that I could dynamically add/remove new workers to a running >>>> standalone Spark cluster by simply triggering: >>>> >>>> start-slave.sh (SPARK_MASTER_ADDR) >>>> >>>> and >>>> >>>> stop-slave.sh >>>> >>>> E.g., I could instantiate a new AWS instance and just add it to a >>>> running cluster without needing to add it to slaves file and restarting the >>>> whole cluster. >>>> It seems that there's no need for me to stop a running cluster. >>>> >>>> Is this a valid way of dynamically resizing a spark cluster (as of now, >>>> I'm not concerned about HDFS)? Or will there be certain unforeseen problems >>>> if nodes are added/removed this way? >>>> >>> >>> >> >