[Rd] AIX testers needed
jasoncbarnhart at msn.com
Mon May 21 20:55:52 CEST 2007
Thanks for responding.
I don't think it's that simple. That's a soft limit, the hard limit
The results of gc() in the original post indicated that R could
utililize more than 32MB of RAM.
My sysadmin had already increased my memory limits prior to my
Just to confirm here are the results with ulimit -m set to unlimited
prior to calling R.
> xx <- matrix(rep(1e+10,1e7),nrow=1e4,ncol=1e3)
> system("ulimit -m")
> tmp.df <- as.data.frame(cbind(xx,xx,xx))
Error: cannot allocate vector of size 228.9 Mb
----- Original Message -----
From: "Hin-Tak Leung" <hin-tak.leung at cimr.cam.ac.uk>
To: "Jason Barnhart" <jasoncbarnhart at msn.com>
Cc: <r-devel at stat.math.ethz.ch>
Sent: Monday, May 21, 2007 11:02 AM
Subject: Re: [Rd] AIX testers needed
> Jason Barnhart wrote:
>> Thank you for responding.
>> I should have added -a on my ulimit command. Here are its results;
>> which I believe are not the limiting factor.
>> %/ > ulimit -a
>> core file size (blocks, -c) 1048575
>> data seg size (kbytes, -d) unlimited
>> file size (blocks, -f) unlimited
>> max memory size (kbytes, -m) 32768
>> open files (-n) 2000
>> pipe size (512 bytes, -p) 64
>> stack size (kbytes, -s) hard
>> cpu time (seconds, -t) unlimited
>> max user processes (-u) 128
>> virtual memory (kbytes, -v) unlimited
> you think max memory = 32768k (or 32MB) is not limiting?
> Please think again...
More information about the R-devel