We use spark without HDFS--in our case, we just use ansible to copy the
spark executors onto all hosts at the same path. We also load and store our
spark data from non-HDFS sources.

On Tue, Oct 21, 2014 at 4:57 AM, Dick Davies <d...@hellooperator.net> wrote:

> I think Spark needs a way to send jobs to/from the workers - the Spark
> distro itself
> will pull down the executor ok, but in my (very basic) tests I got
> stuck without HDFS.
>
> So basically it depends on the framework. I think in Sparks case they
> assume most
> users are migrating from an existing Hadoop deployment, so HDFS is
> sort of assumed.
>
>
> On 20 October 2014 23:18, CCAAT <cc...@tampabay.rr.com> wrote:
> > On 10/20/14 11:46, Steven Schlansker wrote:
> >
> >
> >> We are running Mesos entirely without HDFS with no problems.  We use
> >> Docker to distribute our
> >> application to slave nodes, and keep no state on individual nodes.
> >
> >
> >
> > Background: I'm building up a 3 node cluster to run mesos and spark. No
> > legacy Hadoop needed or wanted. I am using btrfs for the local file
> system,
> > with (2) drives set up for raid1 on each system.
> >
> > So you  are suggesting that I can install mesos + spark + docker
> > and not a DFS on these (3) machines?
> >
> >
> > Will I need any other softwares? My application is a geophysical
> > fluid simulator, so scala, R, and all sorts of advanced math will
> > be required on the cluster for the Finite Element Methods.
> >
> >
> > James
> >
> >
>

Reply via email to