Hi, I'm operating Spark in standalone cluster configuration (3 slaves) and have some question.
1. How can I stop a slave on the specific node? Under `sbin/` directory, there are `start-{all,master,slave,slaves}` and `stop-{all,master,slaves}`, but no `stop-slave`. Are there any way to stop the specific (e.g., the 2nd) slave via command line? 2. How can I check cluster status from command line? Are there any way to confirm that all Master / Workers are up and working without using Web UI? Thanks in advance! -- Kenichi Maehashi --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org