Hi,

> I’m in the ‘big data’ practice at Accenture and am also a big fan of
> Sagemath, using it almost daily.  I’ve recently started wondering what
> interest you and the community would have in integrating Sagemath a
> bit with Hadoop or another map-reduce framework so that operations
> written in Sage could be shipped off to a running cluster for some
> distributed computation, along with being able to run Sage on much
> larger datasets than would fit on a single host.
> 
> There are some existing solutions which are clients to Hadoop’s map-
> reduce framework, such as Hive, which provides a pretty standard SQL-
> like front-end to a live cluster, that made me wonder if there’s not a
> way to extend the usefulness of Sagemath in a similar way.
> 
> I'd love to hear what interest there was in this idea.

I'm interested. But in the practical application I've in mind, the data isn't
there it is generated on the fly by sage itself. The problem it to gather
information (eg: the size) on huge sets from combinatorics. Those sets are
generated by a choice tree. The exploration of the various branches can
obviously be made in parallel.

If you want more information on the problem you can have a look at the
SearchForest [1]. I student made a prototype which can be found at [2].

Cheers,

Florent


[1] http://combinat.sagemath.org/doc/reference/sage/combinat/backtrack.html
[2] http://combinat.sagemath.org/patches/file/931c5a1e58bd/SFparalell-fh.patch

-- 
To post to this group, send an email to sage-devel@googlegroups.com
To unsubscribe from this group, send an email to 
sage-devel+unsubscr...@googlegroups.com
For more options, visit this group at http://groups.google.com/group/sage-devel
URL: http://www.sagemath.org

Reply via email to