[R] Reading huge chunks of data from MySQL into Windows R

bogdan romocea br44114 at yahoo.com
Mon Jun 6 16:26:49 CEST 2005


You don't say what you want to do with the data, how many columns you
have etc. However, I would suggest proceeding in this order:
1. Avoid R; do everything in MySQL.
2. Use random samples.
3. If for some reason you need to process all 160 million rows in R, do
it in a loop. Pull no more than, say, 50-100k rows at a time. This
approach would allow you to process billions of rows without the memory
and disk requirements going through the roof.

hth,
b.


-----Original Message-----
From: Dubravko Dolic [mailto:Dubravko.Dolic at komdat.com]
Sent: Monday, June 06, 2005 9:31 AM
To: r-help at stat.math.ethz.ch
Subject: [R] Reading huge chunks of data from MySQL into Windows R


Dear List,

 

I'm trying to use R under Windows on a huge database in MySQL via ODBC
(technical reasons for this...). Now I want to read tables with some
160.000.000 entries into R. I would be lucky if anyone out there has
some good hints what to consider concerning memory management. I'm not
sure about the best methods reading such huge files into R. for the
moment I spilt the whole table into readable parts stick them together
in R again. 

 

Any hints welcome.

 

 

 

Dubravko Dolic

Statistical Analyst

 

Email: dubravko.dolic at komdat.com


	[[alternative HTML version deleted]]

______________________________________________
R-help at stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide!
http://www.R-project.org/posting-guide.html




		
__________________________________ 

Have fun online with music videos, cool games, IM and more. Check it out!




More information about the R-help mailing list