Dear List, Following the below question I have a question of my own: Suppose that I have large matrices which are produced sequentially and must be used sequentially in the reverse order. I do not have enough memory to store them and so I would like to write them to disk and then read them. This raises two questions: 1) what is the fastest (and the most economic space-wise) way to do this? 2) functions like write, write.table, etc. write the data the way it is printed and this may result in a loss of accuracy. Is there any way to prevent this, except for setting the "digits" option to a higher value or using format prior to writing the data? Is it possible to write binary files (similar to Fortran)?
Any suggestion will be greatly appreciated. --- Wojciech Gryc <[EMAIL PROTECTED]> wrote: > Hi, > > I'm currently working with data that has values as > large as 99,000,000 > but is accurate to 6 decimal places. Unfortunately, > when I load the > data using read.table(), it rounds everything to the > nearest integer. > Is there any way for me to preserve the information > or work with > arbitrarily large floating point numbers? > > Thank you, > Wojciech > > -- > > Five Minutes to Midnight: > Youth on human rights and current affairs > http://www.fiveminutestomidnight.org/ > > ______________________________________________ > R-help@r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide > http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, > reproducible code. > ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.