[R] heap size trouble

Douglas Bates bates at stat.wisc.edu
Wed May 31 15:36:29 CEST 2000


Jim Lemon <bitwrit at ozemail.com.au> writes:

> karamian wrote:
> 
> ...I want to load a file that contains 93 thousand raws and 22 colums of
> data (essentially float)...
> 
> I just had to process over 199000 records with four numeric values.  If
> I remember correctly, I used:
> 
> --vsize 30M  --nsize 500000
> 
> which pretty much ate all the RAM (64M) I had.  Don't forget to "rm" big
> data sets before you exit, or R will bomb when you next try to load
> without the increased memory.  Just reread from the data file when you
> need them again (and it helps to exit other apps before starting R to
> avoid disk thrashing).

Another approach is to use a relational database to store such a large
table and load the table into R from the database.  There are several
interfaces into R from relational databases.
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
Send "info", "help", or "[un]subscribe"
(in the "body", not the subject !)  To: r-help-request at stat.math.ethz.ch
_._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._



More information about the R-help mailing list