[R] Row limit for read.table
Frank McCown
fmccown at cs.odu.edu
Wed Jan 17 16:39:56 CET 2007
I have been trying to read in a large data set using read.table, but
I've only been able to grab the first 50,871 rows of the total 122,269 rows.
> f <-
read.table("http://www.cs.odu.edu/~fmccown/R/Tchange_rates_crawled.dat",
header=TRUE, nrows=123000, comment.char="", sep="\t")
> length(f$change_rate)
[1] 50871
From searching the email archives, I believe this is due to size limits
of a data frame. So...
1) Why doesn't read.table give a proper warning when it doesn't place
every read item into a data frame?
2) Why isn't there a parameter to read.table that allows the user to
specify which columns s/he is interested in? This functionality would
allow extraneous columns to be ignored which would improve memory usage.
I've already made a work-around by loading the table into mysql and
doing a select on the 2 columns I need. I just wonder why the above 2
points aren't implemented. Maybe they are and I'm totally missing it.
Thanks,
Frank
--
Frank McCown
Old Dominion University
http://www.cs.odu.edu/~fmccown/
More information about the R-help
mailing list