[R] Reading in 9.6GB .DAT File - OK with 64-bit R?
sarah.goslee at gmail.com
Fri Mar 9 00:54:29 CET 2012
On Thu, Mar 8, 2012 at 6:45 PM, RHelpPlease <rrumple at trghcsolutions.com> wrote:
> Hi Jeff & Steve,
> Thanks for your responses. After seven hours R/machine ran out of memory
> (and thus ended). Currently the machine has 4GB RAM. I'm looking to
> install more RAM tomorrow.
You can't load a 9.6GB dataset into 4GB of RAM.
> I will look into SQLLite3; thanks!
> I've read that SQL would be a great program for data of this size (read-in,
> manipulate), but I understand there is a hefty price tag (similar to the
> cost of SAS? [licensing]). At this time I'm looking for a low-cost
> solution, if possible. After this data event, a program like SQL would not
> be needed in the future; also, with these multiple data sets to synthesize,
> only a handful are of this size.
sqlite3 is free, and doesn't require you to set up a server and client. More
powerful relational database solutions like MySQL and postgreSQL are
also available for free, but require more effort to configure and use.
> Thanks & please lend any other advice!
More information about the R-help