I have an R script with the following source code:
genofile<-read.table("D_G.txt", header=T, sep=',')
genofile<-genofile[genofile$"GC_SCORE">0.15,]
cat(unique(as.vector(genofile[,2])), file="GF_uniqueIDs.txt", sep='\n')
D_G.txt is a huge file, about 5 GBytes.
Now, the computation is performed on a Microsoft HPC cluster so, as you know, when I submit the job it gets "splitted" across different physical nodes; in my case each one has 4 GB of RAM memory.
Well, after a variable amount of time, I get the infamous error cannot allocate vector of size xxx Mb
message. I've tried to use a switch which limits the usable memory:
--max-memory=1GB
but nothing change.
I've tried Rscript 2.15.0 both 32 and 64 bit with no luck.