On Tue, 30 Oct 2007, David Bickel wrote:
> To help me make choices regarding a platform for running high-memory R
> processes in parallel, I would appreciate any responses to these
> questions:
>
> 1. Does the amount of RAM available to an R session depend on the
> processor (Intel vs. Sun) or on the OS (various Linux distributions vs.
> Solaris)?
Yes.
It depends on whether R uses 64-bit or 32-bit pointers. For 64-bit R you need a
64-bit processor, an operating system that will run 64-bit programs, and a
compiler that will produce them.
I'm not sure what the current Intel offerings are, but you can compile and run
64-bit on AMD Opteron (Linux) and Sun (Solaris) systems.
> 2. Does R have any built-in limitations of RAM available to a session?
> For example, could it make use of 16 GB in one session given the right
> processor/OS platform?
R does have built-in limitations even in a 64-bit system, but they are large.
It is certainly possible to use more than 16Gb of memory.
The main limit is that the length of a vector is stored in a C int, and so is
no more than 2^31-1, or about two billion. A numeric vector of that length
would take up 16Gb on its own.
> 3. Is there anything else I should consider before choosing a processor
> and OS?
I don't think there is anything else R-specific.
-thomas
Thomas Lumley Assoc. Professor, Biostatistics
[EMAIL PROTECTED] University of Washington, Seattle
______________________________________________
[email protected] mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.