[R] error loading huge .RData
Peter Dalgaard BSA
p.dalgaard at biostat.ku.dk
Wed Apr 24 14:39:15 CEST 2002
"Liaw, Andy" <andy_liaw at merck.com> writes:
> I appreciate your comments, and practice everything that you preach.
> However, that workspace image contains only 2~3 R objects: the input and
> output of a single R command. I knew there could be problems, so I've
> stripped it down to the bare minimum. Yes, I also kept the commands in a
> script. That single command (in case you want to know: a random forest run
> with 4000 rows and nearly 7000 variables) took over 3 days to run. There's
> not a whole lot I can do here when the data is so large.
Hmm. You could be running into some sort of situation where data
temporarily take up more space in memory than they need to. It does
sound like a bit of a bug if R can write images that are bigger than
it can read. Not sure how to proceed though. Does anyone on R-core
have a similarly big system and a spare gigabyte of disk? Is it
possible to create a mock-up of similarly organized data that displays
the same effect, but takes less than three days?
BTW: Did we ever hear what system this is happening on?
O__ ---- Peter Dalgaard Blegdamsvej 3
c/ /'_ --- Dept. of Biostatistics 2200 Cph. N
(*) \(*) -- University of Copenhagen Denmark Ph: (+45) 35327918
~~~~~~~~~~ - (p.dalgaard at biostat.ku.dk) FAX: (+45) 35327907
r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
Send "info", "help", or "[un]subscribe"
(in the "body", not the subject !) To: r-help-request at stat.math.ethz.ch
More information about the R-help