In R, I'm trying to read in a basic CSV file of about 42,900 rows (confirmed by Unix's wc -l). The relevant code is
vecs <- read.csv("feature_vectors.txt", header=FALSE, nrows=50000)
where nrows is a slight overestimate because why not. However,
>> dim(vecs)
[1] 16853 5
indicating that the resultant data frame has on the order of 17,000 rows. Is this a memory issue? Each row consists of a ~30 character hash code, a ~30 character string, and 3 integers, so the total size of the file is only about 4MB.
If it's relevant, I should also note that a lot of the rows have missing fields.
Thanks for your help!