[Rd] working with huge memory: single precision?
simon.urbanek at r-project.org
Wed Aug 27 15:43:16 CEST 2014
On Aug 27, 2014, at 4:03 AM, Mario Emmenlauer <mario at emmenlauer.de> wrote:
> I'm very new to R and don't know much about it yet. I would like
> to develop R-programs that work with data of sizes of 10^10 - 10^11
> data points. We have very-high-memory machines with ~256 GB, but it
> would significantly help if I could store the data points in single
> precision in RAM instead of double precision. Is that possible?
You can (e.g. in raw vectors), but it may not help much since you can't operate on them directly, since no functions in R know how to deal with single-precision floats - all arithmetics are on double precision vectors. If you want to load the data in memory but only work on small pieces, then it would work since you could extract the piece, convert to doubles and carry on.
> In the documentation I found a sentence saying its not supported,
> at least not out of the box. But I am quite desperate and would also
> consider working with an alpha version or with extension packages?
> Ideally I would like type promotion to work, i.e. that when using
> the data in math operations they should be promoted to double.
That won't work automatically that way, but you cloud write methods for operators on your new type class and implement it as coercion + call to the regular operators. You may take a hint from the 64-bit int packages and I dimly recall that some of the mem-mapping packages (bigMemory, ff, ..) may also support single-precision storage.
> Any help is greatly appreciated! All the best,
> Mario Emmenlauer BioDataAnalysis Mobil: +49-(0)151-68108489
> Balanstrasse 43 mailto: mario.emmenlauer * unibas.ch
> D-81669 München http://www.marioemmenlauer.de/
> R-devel at r-project.org mailing list
More information about the R-devel