[R] FW: Large datasets in R
Marshall Feldman
marsh at uri.edu
Tue Jul 18 15:20:06 CEST 2006
Hi,
I have two further comments/questions about large datasets in R.
1. Does R's ability to handle large datasets depend on the operating
system's use of virtual memory? In theory, at least, VM should make the
difference between installed RAM and virtual memory on a hard drive
primarily a determinant of how fast R will calculate rather than whether or
not it can do the calculations. However, if R has some low-level routines
that have to be memory resident and use more memory as the amount of data
increases, this may not hold. Can someone shed light on this?
2. Is What 64-bit versions of R are available at present?
Marsh Feldman
The University of Rhode Island
-----Original Message-----
From: Thomas Lumley [mailto:tlumley at u.washington.edu]
Sent: Monday, July 17, 2006 3:21 PM
To: Deepankar Basu
Cc: r-help at stat.math.ethz.ch
Subject: Re: [R] Large datasets in R
On Mon, 17 Jul 2006, Deepankar Basu wrote:
> Hi!
>
> I am a student of economics and currently do most of my statistical work
> using STATA. For various reasons (not least of which is an aversion for
> proprietary software), I am thinking of shifting to R. At the current
> juncture my concern is the following: would I be able to work on
> relatively large data-sets using R? For instance, I am currently working
> on a data-set which is about 350MB in size. Would be possible to work
> data-sets of such sizes using R?
The answer depends on a lot of things, but most importantly
1) What you are going to do with the data
2) Whether you have a 32-bit or 64-bit version of R
3) How much memory your computer has.
In a 32-bit version of R (where R will not be allowed to address more than
2-3Gb of memory) an object of size 350Mb is large enough to cause problems
(see eg the R Installation and Adminstration Guide).
If your 350Mb data set has lots of variables and you only use a few at a
time then you may not have any trouble even on a 32-bit system once you
have read in the data.
If you have a 64-bit version of R and a few Gb of memory then there should
be no real difficulty in working with that size of data set for most
analyses. You might come across some analyses (eg some cluster analysis
functions) that use n^2 memory for n observations and so break down.
-thomas
Thomas Lumley Assoc. Professor, Biostatistics
tlumley at u.washington.edu University of Washington, Seattle
More information about the R-help
mailing list