[R] Reading in 9.6GB .DAT File - OK with 64-bit R?
rrumple at trghcsolutions.com
Fri Mar 9 00:45:25 CET 2012
Hi Jeff & Steve,
Thanks for your responses. After seven hours R/machine ran out of memory
(and thus ended). Currently the machine has 4GB RAM. I'm looking to
install more RAM tomorrow.
I will look into SQLLite3; thanks!
I've read that SQL would be a great program for data of this size (read-in,
manipulate), but I understand there is a hefty price tag (similar to the
cost of SAS? [licensing]). At this time I'm looking for a low-cost
solution, if possible. After this data event, a program like SQL would not
be needed in the future; also, with these multiple data sets to synthesize,
only a handful are of this size.
Thanks & please lend any other advice!
View this message in context: http://r.789695.n4.nabble.com/Reading-in-9-6GB-DAT-File-OK-with-64-bit-R-tp4457220p4458042.html
Sent from the R help mailing list archive at Nabble.com.
More information about the R-help