My problem is simple calculations on large data sets (about 25 million rows and 10 columns, i.e. aroung 1GB data). My system:
32bits/Windows7/4Gb Ram/R Studio 0.96, R 2.15.2
I can link to my database using the BigMemory package. And use the functions above my db. I can also do this with the ff package, filehash, etc.
The problem is the calculation of simple calculations (in the form of unique values, means, etc.). I have a typical problem
"cannot select vector size n mb"
where n can be from 70 to 95 mb in size, etc.
I know of all the (I think) solutions provided so far about this:
increase RAM. launch R with inline code "--max-mem-size XXXX", use memory.limit() and memory-size() commands, use rm() and gc(), work on 64bit, close other programs, free memory, reboot, use packages bigmemory, ff, filehash, sql, etc etc. improve your data, use integers, shorts, etc. ... check memory usage of intermediate calculations, ... etc.
All this has been checked, completed (with the exception of switching to another system / machine, kindly), etc.
But I still get that they "cannot allocate a vector size of n mb", where n is about 90 mb, for example, practically without using memory from R or other programs, all this restarts, updates ... I know the differences between free memory and distribution from windows and R etc. but
This makes no sense because the available memory is more than 3 GB. I suspect the reason is that it is related to Windows32b-R memory management, but it seems like it's almost a joke to buy 4 GB of RAM or switch the entire system to 64 bits to allocate 70 MB.
Is there something I am missing?