Mag Gam wrote:
Hi John:
Well, we run a lot of statistical analysis and our code loads a lot of data into a vector for fast calculations. I am not sure how else to do these calculations fast without loading it into memory. Thats why we have to do it this way.
well, if you got several processes that each need 32GB in a 64GB machine, you're gonna end up swapping.
the traditional way of doing this sort of thing on limited memory machines was to take a sequential pass through the data, calculating the statistics on the fly. I know that kind of thing is very difficult for some algorithms (FFT's are notorious for being unfriendly to sequential processing), but for many algorithms, a few sequential passes of calculations can be /faster/ than random access and swapping when theres memory/process contention.