[R] R memory usage

Jun Ding dingjun_cn at yahoo.com
Thu Aug 9 03:08:27 CEST 2007

Hi All, 

I have two questions in terms of the memory usage in R
(sorry if the questions are naive, I am not familiar
with this at all). 

1) I am running R in a linux cluster. By reading the R
helps, it seems there are no default upper limits for
vsize or nsize. Is this right? Is there an upper limit
for whole memory usage? How can I know the default in
my specific linux environment? And can I increase the

2) I use R to read in several big files (~200Mb each),
and then I run:


I get:

            used  (Mb) gc trigger   (Mb)  max used   
Ncells  23083130 616.4   51411332 1372.9  51411332 
Vcells 106644603 813.7  240815267 1837.3 227550003 


What do columns of "used", "gc trigger" and "max used"
mean? It seems to me I have used 616Mb of Ncells and
813.7Mb of Vcells. Comparing with the numbers of "max
used", I still should have enough memory. But when I

object.size(area.results)   ## area.results is a big

I get an error message:

Error: cannot allocate vector of size 32768 Kb

Why is that? Looks like I am running out of memory. Is
there a way to solve this problem?

Thank you very much!



More information about the R-help mailing list