[R] for loop performance
p.pagel at wzw.tum.de
Thu Apr 14 15:14:16 CEST 2011
On Thu, Apr 14, 2011 at 06:50:56AM -0500, Barth B. Riley wrote:
> Thank you Phillip for your post. I am reading in:
> 1. a 3 x 100 item parameter file (floating point and integer data)
> 2. a 100 x 1000 item response file (integer data)
> 3. a 6 x 1000 person parameter file (contains simulation condition
> information, person measures)
> 4. I am then computing several statistics used in subsequent ROC
> analyses, the AUCs being stored in a 6000 x 15 matrix of floating
> point numbers
> I am using read.table for #1-#3 and write.table for #4. The process
> of reading files (#1-#3) and writing to file is done over 6,000
A few ideas:
1) try to use the colClasses argument to read.table. That way R will
not have to guess the data type of columns.
2) When you say 6000 iterations - do you mean you are reading/writing the SAME
files over and over again? Or do you have 6000 sets of files? In the
former case the obvious advice would be to only read them once.
3) If the input files were generated in R, another option would be to
save()/load() them rather than using write.table()/read.table().
4) If the came from some other application, possibly storing
everything in a database may speed up things.
5) Is your data on a file server? If yes: try moving it to the local
disc temporarily to see if network i/o is limiting your speed.
6) Whatever you try to improve performance - measure the effects
rather than rely on your impression (system.time, Rprof, ...) in order
to find out what part of the program is actually eating up the most
Dr. Philipp Pagel
Lehrstuhl für Genomorientierte Bioinformatik
Technische Universität München
85354 Freising, Germany
More information about the R-help