Dear R experts:  I want to learn what the practically useful memory
limits are for good work with R.

(My specific problem is that I want work with daily stock returns.
In ASCII, the data set is about 72 million returns, that would have to
go into a sparse matrix (not all stocks exist for the whole series).
As a guess, this will consume about 700MB.  My main use will be linear
operations---regressions, means, etc.)

I am on linux, so I can create swap space, but I am concerned that the
thrashing will be so bad that the computer will become worthless.  In
fact, the last time I used it was over 3 years ago.  Since then, I
have just turned it off.

I have 2GB of RAM right now, and could upgrade this to 4GB.

Are there some general guidelines as to what the relationship between
data sets and memory should be under R?  I know this will vary with
the task involved, but some guidance would be better than none.

regards,

/iaw

______________________________________________
[email protected] mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to