[R] What is the largest in memory data object you've worked with in R?

Joris Meys jorismeys at gmail.com
Sat Jun 5 23:04:46 CEST 2010

You have to take some things into account :
- the maximum memory set for R might not be the maximum memory available
- R needs the memory not only for the dataset. Matrix manipulations
require frquently double of the amount of memory taken by the dataset.
- memory allocation is important when dealing with large datasets.
There is plenty of information about that
- R has some packages to get around memory problems with big datasets.

Read this discussione for example :

and this page of Matthew Keller is a good summary too :


On Sat, Jun 5, 2010 at 12:32 AM, Nathan Stephens <nwstephens at gmail.com> wrote:
> For me, I've found that I can easily work with 1 GB datasets.  This includes
> linear models and aggregations.  Working with 5 GB becomes cumbersome.
> Anything over that, and R croaks.  I'm using a dual quad core Dell with 48
> GB of RAM.
> I'm wondering if there is anyone out there running jobs in the 100 GB
> range.  If so, what does your hardware look like?
> --Nathan
>        [[alternative HTML version deleted]]
> ______________________________________________
> R-help at r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-help
> PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
> and provide commented, minimal, self-contained, reproducible code.

Ghent University
Faculty of Bioscience Engineering
Department of Applied mathematics, biometrics and process control

tel : +32 9 264 59 87
Joris.Meys at Ugent.be
Disclaimer : http://helpdesk.ugent.be/e-maildisclaimer.php

More information about the R-help mailing list