[R] Absolute ceiling on R's memory usage = 4 gigabytes?
laurent.buffat at it-omics.com
Tue Jul 13 14:35:30 CEST 2004
I am very surprise by your answers :
When I try to make an affybatch with bioconductor and R 1.9.1, I was unable
to read and normalise more than 80 HU-133A CEL file with a Linux 32 bits
computer and 4 GB of RAM + 8 GB of swap (Of course, without any other
process on the computer and I don't' want do to "JustRMA" because I want the
probe level information in the affybatch, And it's not a limit in R
configuration, because if I follow the memory usage during the R session, R
is using all the 4GB RAM memory (the swap is not use) before the memory
For this raison we are planning to buy a 64 bits under Linux, but, if with
Mac OS X and 1.5 GB of RAM, we can solve this problem, I will buy a Mac and
not a linux 64 bits computer.
So, what kind of normalization are you doing ? Some one with bioconductor
and the affy package or an other one ? Could you precise ?
For the other R & BioC :
Do you think that there is a difference between linux and MacOS for the
memory management under R ?
What is a "good" hardward solution for "R / Linux 64 bits" ?
Thanks for your help.
De : r-help-bounces at stat.math.ethz.ch
[mailto:r-help-bounces at stat.math.ethz.ch] De la part de Tae-Hoon Chung
Envoye : vendredi 2 juillet 2004 01:52
A : Kort, Eric
Cc : r-help at stat.math.ethz.ch
Objet : Re: [R] Absolute ceiling on R's memory usage = 4 gigabytes?
It seems a little bit puzzling to me. Which Affymetrix chip do you use?
The reason I'm asking this is that yesterday I was able to normalize
150 HU-133A CEL files (containing 22283 probes) using R 1.9.1 in Mac OS
X 10.3.3 with 1.5 GB memory. If your chip has more probes than this,
then it must be understandable ...
On Jul 1, 2004, at 2:59 PM, Kort, Eric wrote:
> Hello. By way of background, I am running out of memory when
> attempting to normalize the data from 160 affymetrix microarrays using
> justRMA (from the affy package). This is despite making 6 gigabytes
> of swap space available on our sgi irix machine (which has 2 gigabytes
> of ram). I have seen in various discussions statements such as "you
> will need at least 6 gigabytes of memory to normalize that many
> chips", but my question is this:
> I cannot set the memory limits of R (1.9.1) higher than 4 gigabytes as
> attempting to do so results in this message:
> WARNING: --max-vsize=4098M=4098`M': too large and ignored
> I experience this both on my windows box (on which I cannot allocate
> more than 4 gigabytes of swap space anyway), and on an the above
> mentioned sgi irix machine (on which I can). In view of that, I do
> not see what good it does to make > 4 gigabytes of ram+swap space
> available. Does this mean 4 gigabytes is the absolute upper limit of
> R's memory usage...or perhaps 8 gigabytes since you can set both the
> stack and the heap size to 4 gigabytes?
> This email message, including any attachments, is for the
> R-help at stat.math.ethz.ch mailing list
> PLEASE do read the posting guide!
Tae-Hoon Chung, Ph.D
Post-doctoral Research Fellow
Molecular Diagnostics and Target Validation Division
Translational Genomics Research Institute
1275 W Washington St, Tempe AZ 85281 USA
R-help at stat.math.ethz.ch mailing list
PLEASE do read the posting guide!
More information about the R-help