Re: [R] Large dataset + randomForest

2007-07-27 Thread Florian Nigsch
x 150) and is diagonal. That might not solve it, but it should help. Max -Original Message- From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Florian Nigsch Sent: Thursday, July 26, 2007 2:07 PM To: r-help@stat.math.ethz.ch Subject: [R] Large dataset

Re: [R] Large dataset + randomForest

2007-07-27 Thread Florian Nigsch
I compiled the newest R version on a Redhat Linux (uname -a = Linux .cam.ac.uk 2.4.21-50.ELsmp #1 SMP Tue May 8 17:18:29 EDT 2007 i686 i686 i386 GNU/Linux) with 4GB of physical memory. The step when the whole script crashed is within the randomForest() routine, I do know that because I

[R] Large dataset + randomForest

2007-07-26 Thread Florian Nigsch
[Please CC me in any replies as I am not currently subscribed to the list. Thanks!] Dear all, I did a bit of searching on the question of large datasets but did not come to a definite conclusion. What I am trying to do is the following: I want to read in a dataset with approx. 100 000 rows

Re: [R] Large dataset + randomForest

2007-07-26 Thread Kuhn, Max
@stat.math.ethz.ch Subject: [R] Large dataset + randomForest [Please CC me in any replies as I am not currently subscribed to the list. Thanks!] Dear all, I did a bit of searching on the question of large datasets but did not come to a definite conclusion. What I am trying to do

Re: [R] large dataset!

2006-07-03 Thread Anupam Tyagi
JENNIFER HILL jh1030 at columbia.edu writes: Hi, I need to analyze data that has 3.5 million observations and about 60 variables and I was planning on using R to do this but I can't even seem to read in the data. It just freezes and ties up the whole system -- and this is on a Linux box

Re: [R] large dataset!

2006-07-03 Thread rdporto1
at http://finzi.psych.upenn.edu/R/Rhelp02a/archive/76583.html HTH, Rogerio Porto. -- Cabeçalho original --- De: [EMAIL PROTECTED] Para: r-help@stat.math.ethz.ch Cópia: Data: Sun, 2 Jul 2006 10:12:25 -0400 (EDT) Assunto: [R] large dataset! Hi, I need to analyze data that has

[R] large dataset!

2006-07-02 Thread JENNIFER HILL
Hi, I need to analyze data that has 3.5 million observations and about 60 variables and I was planning on using R to do this but I can't even seem to read in the data. It just freezes and ties up the whole system -- and this is on a Linux box purchased about 6 months ago on a dual-processor PC

Re: [R] large dataset!

2006-07-02 Thread Andrew Robinson
Jennifer, it sounds like that's too much data for R to hold in your computer's RAM. You should give serious consideration as to whether you need all those data for the models that you're fitting, and if so, whether you need to do them all at once. If not, think about pre-processing steps, using

Re: [R] large dataset!

2006-07-02 Thread miguel manese
Hello Jennifer, I'm writing a package SQLiteDF for Google SOC2006, under the supervision of Prof. Bates Prof. Riley. Basically, it stores data frame into sqlite databases (i.e. in a file) and aims to be transparently accessible to R using the same operators for ordinary data frames. Right now,

[R] large dataset import, aggregation and reshape

2005-04-24 Thread Christoph Lehmann
Dear useRs We have a data-set (comma delimited) with 12Millions of rows, and 5 columns (in fact many more, but we need only 4 of them): id, factor 'a' (5 levels), factor 'b' (15 levels), date-stamp, numeric measurement. We run R on suse-linux 9.1 with 2GB RAM, (and a 3.5GB swap file). on

Re: [R] large dataset import, aggregation and reshape

2005-04-24 Thread Renaud Lancelot
Christoph Lehmann a écrit : Dear useRs We have a data-set (comma delimited) with 12Millions of rows, and 5 columns (in fact many more, but we need only 4 of them): id, factor 'a' (5 levels), factor 'b' (15 levels), date-stamp, numeric measurement. We run R on suse-linux 9.1 with 2GB RAM, (and a