[R] memory problem on Suse

Marc Schwartz marc_schwartz at me.com
Fri Dec 11 17:02:24 CET 2009


On Dec 11, 2009, at 6:24 AM, Ambrosi Alessandro wrote:

>
> Dear all, I am meeting some problems with  memory allocation. I know  
> it is an old issue, I'm sorry.
> I looked for a solution in the FAQs and manuals, mails, but without  
> finding the working answer.
> I really hope you can help me.
> For instance, if I try to read micorarray data I get:
>
>> mab=ReadAffy(cdfname="hgu133plus2cdf")
> Error: cannot allocate vector of size 858.0 Mb
>>
>
> I get similar errors with smaller objects, smaller data sets or  
> other procedures
> ("Error: cannot allocate vector of size 123.0 Mb").
> I'm running R with Suse 11.1 Linux OS, on two Xeon processors (8  
> cores), 32 GB RAM.
> I suppose I have enough resources to manage these objects and data  
> files....
>
> Any suggestions or hints will be really appreciated!
> Many thanks in advance.
> Alessandro

Well, you are running into a situation where there is not a contiguous  
chunk of RAM available in the sizes referenced, for allocation to the  
vector.

Presuming that you are running a 64 bit version of SUSE (what does  
'uname -a' show in a system console), you should also check to be sure  
that you are also running a 64 bit version of R. What does:

   .Machine$sizeof.pointer

show?

If it returns 4, then you are running a 32 bit version of R, which  
cannot take advantage of your 64 bit platform. You should install a 64  
bit version of R.

HTH,

Marc Schwartz




More information about the R-help mailing list