[R] Boundaries of R
Ben Bolker
bbolker at gmail.com
Fri Feb 18 15:42:22 CET 2011
Mike Marchywka <marchywka <at> hotmail.com> writes:
> Duncan Murdoch:
> > Vectors are limited to about 2 billion entries (2^31 - 1). Matrices are
> > vectors, so that limit applies to the total count of entries.
> > Dataframes are lists of vectors, so that limit applies separately to the
> > numbers of rows and columns.
> >
> > Simple R code keeps everything in memory, so you're likely to run into
> > hardware limits if you start working with really big vectors. There are
> > a number of packages that alleviate that by paging data in and out, but
> > it takes a bit of work on your part to use them. As far as I know,
>
> Do you have more details here?
[snip]
The best starting point is the "high performance computing" task
view on CRAN, which gives an (as far as I know) up-to-date description
of the various packages that are available for handing large/out-of-memory
data sets. Some of these are RDBMS interfaces, some are systems for
file-backed objects, some are large/out-of-memory algorithms such as
lm or glm for big objects.
More information about the R-help
mailing list