Hi John: Well, we run a lot of statistical analysis and our code loads a lot of data into a vector for fast calculations. I am not sure how else to do these calculations fast without loading it into memory. Thats why we have to do it this way. TIA On Thu, Oct 16, 2008 at 1:00 PM, John R Pierce <pierce at hogranch.com> wrote: > Mag Gam wrote: >> >> Hello All: >> >> Running 5.2 at our university. We have several student's processes >> that take up too much memory. Our system have 64G of RAM and some >> processes take close to 32-48G of RAM. This is causing many problems >> for others. I was wondering if there is a way to restrict memory usage >> per process? If the process goes over 32G simply kill it. Any thoughts >> or ideas? >> >> > > In /etc/profile, use "ulimit -v NNNN" (in kilobytes) to limit the max > virtual of all processes spawned by that shell > > > 32G per process on a 64G machine sounds like a bit much. wouldn't a limit > more like 4GB per user session be more appropriate on a multiuser system? > _______________________________________________ > CentOS mailing list > CentOS at centos.org > http://lists.centos.org/mailman/listinfo/centos >