Thank you so much Steve.

The computer I'm currently working with is a 32 bit windows 7 OS. And RAM
is only 4GB so I guess thats a big limitation.
El 18/08/2013 03:11, "Steve Lianoglou" <lianoglou.st...@gene.com> escribió:

> Hi Paul,
>
> On Sun, Aug 18, 2013 at 12:56 AM, Paul Bernal <paulberna...@gmail.com>
> wrote:
> > Thanks a lot for the valuable information.
> >
> > Now my question would necessarily be, how many columns can R handle,
> > provided that I have millions of rows and, in general, whats the maximum
> > amount of rows and columns that R can effortlessly handle?
>
> This is all determined by your RAM.
>
> Prior to R-3.0, R could only handle vectors of length 2^31 - 1. If you
> were working with a matrix, that meant that you could only have that
> many elements in the entire matrix.
>
> If you were working with a data.frame, you could have data.frames with
> 2^31-1 rows, and I guess as many columns, since data.frames are really
> a list of vectors, the entire thing doesn't have to be in one
> contiguous block (and addressable that way)
>
> R-3.0 introduced "Long Vectors" (search for that section in the release
> notes):
>
> https://stat.ethz.ch/pipermail/r-announce/2013/000561.html
>
> It almost doubles the size of a vector that R can handle (assuming you
> are running 64bit). So, if you've got the RAM, you can have a
> data.frame/data.table w/ billion(s) of rows, in theory.
>
> To figure out how much data you can handle on your machine, you need
> to know the size of real/integer/whatever and the number of elements
> of those you will have so you can calculate the amount of RAM you need
> to load it all up.
>
> Lastly, I should mention there are packages that let you work with
> "out of memory" data, like bigmemory, biglm, ff. Look at the HPC Task
> view for more info along those lines:
>
> http://cran.r-project.org/web/views/HighPerformanceComputing.html
>
>
> >
> > Best regards and again thank you for the help,
> >
> > Paul
> > El 18/08/2013 02:35, "Steve Lianoglou" <lianoglou.st...@gene.com>
> escribió:
> >
> >> Hi Paul,
> >>
> >> First: please keep your replies on list (use reply-all when replying
> >> to R-help lists) so that others can help but also the lists can be
> >> used as a resource for others.
> >>
> >> Now:
> >>
> >> On Aug 18, 2013, at 12:20 AM, Paul Bernal <paulberna...@gmail.com>
> wrote:
> >>
> >> > Can R really handle millions of rows of data?
> >>
> >> Yup.
> >>
> >> > I thought it was not possible.
> >>
> >> Surprise :-)
> >>
> >> As I type, I'm working with a ~5.5 million row data.table pretty
> >> effortlessly.
> >>
> >> Columns matter too, of course -- RAM is RAM, after all and you've got
> >> to be able to fit the whole thing into it if you want to use
> >> data.table. Once loaded, though, data.table enables one to do
> >> split/apply/combine calculations over these data quite efficiently.
> >> The first time I used it, I was honestly blown away.
> >>
> >> If you find yourself wanting to work with such data, you could do
> >> worse than read through data.table's vignette and FAQ and give it a
> >> spin.
> >>
> >> HTH,
> >>
> >> -steve
> >>
> >> --
> >> Steve Lianoglou
> >> Computational Biologist
> >> Bioinformatics and Computational Biology
> >> Genentech
> >>
> >
> >         [[alternative HTML version deleted]]
> >
> >
> > ______________________________________________
> > R-help@r-project.org mailing list
> > https://stat.ethz.ch/mailman/listinfo/r-help
> > PLEASE do read the posting guide
> http://www.R-project.org/posting-guide.html
> > and provide commented, minimal, self-contained, reproducible code.
> >
>
>
>
> --
> Steve Lianoglou
> Computational Biologist
> Bioinformatics and Computational Biology
> Genentech
>

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to