Hi Pietro,

Thanks for the suggestion, I will have a look at the documentation.

Paulo

On Fri, Feb 13, 2015 at 10:09 AM, Pietro <peter.z...@gmail.com> wrote:

> Dear Paulo,
>
> On Fri, Feb 13, 2015 at 9:57 AM, Paulo van Breugel
> <p.vanbreu...@gmail.com> wrote:
> > I guess this is because the calculations are done in-memory? Any way to
> > avoid this memory problem when using large data sets (something like
> working
> > with memmap objects?)
>
> With memmap you still have a limits of 2Gb I guess, you should try: dask
>
> Dask Array implements the NumPy ndarray interface using blocked
> algorithms, cutting up the large array into many small arrays. This
> lets us compute on arrays larger than memory using all of our cores.
> We coordinate these blocked algorithms using dask graphs.
>
> http://dask.readthedocs.org/en/latest/array.html
>
> I didn't have a chance to try it yet, but it support a numpy array
> syntax, and since you are using quite basic functionalities I think
> you should be able to work with it.
>
> All the best
>
> Pietro
>
_______________________________________________
grass-dev mailing list
grass-dev@lists.osgeo.org
http://lists.osgeo.org/mailman/listinfo/grass-dev

Reply via email to