[R-sig-Geo] memory Usage setting
elw at stderr.org
elw at stderr.org
Tue Sep 11 14:20:39 CEST 2007
> These days in GIS on may have to manipulate big datasets or arrays.
>
> Here I am on WINDOWS I have a 4Gb
> my aim was to have an array of dim 298249 12 10 22 but that's 2.9Gb
It used to be (maybe still is?) the case that a single process could only
'claim' a chunk of max size 2GB on Windows.
Also remember to compute overhead for R objects... 58 bytes per object, I
think it is.
> It is also strange that once a dd needed 300.4Mb and then 600.7Mb (?) as
> also I made some room in removing ZZ?
Approximately double size - many things the interpreter does involve
making an additional copy of the data and then working with *that*. This
might be happening here, though I didn't read your code carefully enough
to be able to be certain.
> which I don't really know if it took into account as the limit is
> greater than the physical RAM of 4GB. ...?
:)
> would it be easier using Linux ?
possibly a little bit - on a linux machine you can at least run a PAE
kernel (giving you a lot more address space to work with) and have the
ability to turn on a bit more virtual memory.
usually with data of the size you're trying to work with, i try to find a
way to preprocess the data a bit more before i apply R's tools to it.
sometimes we stick it into a database (postgres) and select out the bits
we want our inferences to be sourced from. ;)
it might be simplest to just hunt up a machine with 8 or 16GB of memory in
it, and run those bits of the analysis that really need memory on that
machine...
--e
More information about the R-sig-Geo
mailing list