Hey all,

I'm doing some work with machine learning on R (I'm a fairly new user of R),
and I have a question about generating new tables from existing tables.  I'm
currently using a table of measurements I read in from a CSV file to
generate training and validation data set tables for future use in a machine
learning algorithm using the code:

#generate probabilities to divide up training / validation data sets
randomly
device_Prob_Vector <- runif(num_Devices)

#NULL-initialize training and validation sets.  This seems like a bit of a
hack...
training_Set <- measurements[0]
validation_Set <- measurements[0]

#divide up the training and validation data sets from measurements.
for ( i in 1:num_Devices)
{
        if ( device_Prob_Vector[i] > 0.5 )
        {
                training_Set <- rbind(training_Set, measurements[i,])
        }
        else
        {
                validation_Set <- rbind(validation_Set, measurements[i,])
        }
}

This code works correctly, but takes quite a long time to execute.  I
suspect this is because rbind() is dynamically resizing the tables as it
adds new rows to each table of data.  Is there a way to pre-allocate memory
for each of the two tables, and then shrink them after the loop has
completed?  Thanks for the help.

~Nate
-- 
View this message in context: 
http://www.nabble.com/Simple-table-generation-question-tf4056042.html#a11521582
Sent from the R help mailing list archive at Nabble.com.

______________________________________________
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to