Hi all, I have a cluster that has multiple nodes, and the data partition is unified, therefore all my nodes in my computer can access to the data I am working on. Right now, I run Spark in a single node, and it work beautifully.
My question is, Is it possible to run Spark using multiple compute nodes (as a standalone mode, I don't have HDFS/Hadoop installed)? If so, what do I have to add/change to my Spark version or Spark script (either python or scala)? Thanks, Jose -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Multiple-compute-nodes-in-standalone-mode-tp27218.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org