[R-sig-hpc] communicating memory requirments

Martin Morgan mtmorgan at fhcrc.org
Tue Feb 19 19:11:53 CET 2013


On 02/19/2013 05:16 AM, Max Kuhn wrote:
> I have some scripts for a book that I will be publishing and they use
> parallel processing (via foreach). Some of the analyses use more memory
> than some users will have on hand and, as the number of workers increases,
> so do the memory demands.
>
> I'd like to report the memory requirements in a way that most people will
> understand (including me).

not really answering your question, but it seems like parallel evaluation in 
shared memory computers comes with an implicit need to manage memory, and that 
one would rather strive to implement algorithms that do their job in a memory 
efficient way. Probably this means iterating through data and aggregating 
results, which is the approach of biglm. The user shouldn't really be exposed to 
the need to choose a computer (or package) based on memory consumption of 
algorithms. I'm not throwing stones, having seldom managed this myself.

Martin

>
> On OS X, I've run the scripts and did a roiling append of 'top' to capture
> the memory used by the master process and the workers over time.
>
> Can anyone suggest which parameters I should report (e.g. VSIZE)? Is the
> situation appreciably different on Windows?
>
> I admit to being fairly ignorant on this (complicated) subject so any
> approach to informing the users would be very welcome.
>
> Thanks,
>
> Max
>
> 	[[alternative HTML version deleted]]
>
> _______________________________________________
> R-sig-hpc mailing list
> R-sig-hpc at r-project.org
> https://stat.ethz.ch/mailman/listinfo/r-sig-hpc
>


-- 
Computational Biology / Fred Hutchinson Cancer Research Center
1100 Fairview Ave. N.
PO Box 19024 Seattle, WA 98109

Location: Arnold Building M1 B861
Phone: (206) 667-2793



More information about the R-sig-hpc mailing list