i have a mesos cluster, which i deploy spark to by using instructions on http://spark.apache.org/docs/0.7.2/running-on-mesos.html
after that the spark shell starts up fine. then i try the following on the shell: val data = 1 to 10000 val distData = sc.parallelize(data) distData.filter(_< 10).collect() open spark web ui at host:4040 and see an active job. NOW, how do i start workers or spark workers on mesos ? who completes my job? thanks, -- Ani