Re: How production un-ready are Mesos Cassandra, Spark and Kafka Frameworks?

2015-10-17 Thread Sam Bessalah
Thats' really depends on what you're doing. I've been running Spark in production with Mesos for as far as Spark ever got open sourced. Earlier this year, we added Cassandra in the mix by running it through Docker and Marathon in host network mode with volume. Nothing fancy, since it was for non cr

Re: Spark on Mesos / Executor Memory

2015-10-17 Thread David Greenberg
Make sure to check out Cook--this is the exact reason we built it! I gave a talk on it at Mesoscon Europe, so that'll be available online soon :-) On Sat, Oct 17, 2015 at 1:40 PM Bharath Ravi Kumar wrote: > To be precise, the MesosExecutorBackend's Xms & Xmx equal > spark.executor.memory. So ther

Re: Spark on Mesos / Executor Memory

2015-10-17 Thread Bharath Ravi Kumar
To be precise, the MesosExecutorBackend's Xms & Xmx equal spark.executor.memory. So there's no question of expanding or contracting the memory held by the executor. On Sat, Oct 17, 2015 at 5:38 PM, Bharath Ravi Kumar wrote: > David, Tom, > > Thanks for the explanation. This confirms my suspicion

Re: Spark on Mesos / Executor Memory

2015-10-17 Thread Bharath Ravi Kumar
David, Tom, Thanks for the explanation. This confirms my suspicion that the executor was holding on to memory regardless of tasks in execution once it expands to occupy memory in keeping with spark.executor.memory. There certainly is scope for improvement here, though I realize there will substan

Re: Spark on Mesos / Executor Memory

2015-10-17 Thread Tom Arnfeld
Hi Bharath, When running jobs in fine grained mode, each Spark task is sent to mesos as a task which allows the offers system to maintain fairness between different spark application (as you've described). Having said that, unless your memory per-node is hugely undersubscribed when running thes

Re: Spark on Mesos / Executor Memory

2015-10-17 Thread David Greenberg
Spark doesn't automatically cooperate with other frameworks on the cluster. Have a look at Cook (github.com/twosigma/cook) for a spark scheduler on Mesos that is able to react to changing cluster conditions and will scale down the low priority jobs as more high priority ones appear. On Sat, Oct 17,