Not sure what did you aim to solve. When you mention Spark Master, I guess you 
probably mean spark standalone mode? In that case spark cluster does not 
necessary coupled with hadoop cluster. While if you aim to achieve better data 
locality , then yes, run spark worker on HDFS data node might help. And for 
spark Master, I think that doesn't matter much.

Best Regards,
Raymond Liu

-----Original Message-----
From: mharwida [mailto:majdharw...@yahoo.com] 
Sent: Tuesday, January 21, 2014 2:14 AM
To: user@spark.incubator.apache.org
Subject: Spark Master on Hadoop Job Tracker?

Hi,

Should the Spark Master run on the Hadoop Job Tracker node (and Spark workers 
on Task Trackers) or the placement of the Spark Master could reside on any 
Hadoop node?

Thanks
Majd



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Spark-Master-on-Hadoop-Job-Tracker-tp680.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to