[R] Enormous Datasets

2004-11-18 Thread Thomas W Volscho
Dear List, I have some projects where I use enormous datasets. For instance, the 5% PUMS microdata from the Census Bureau. After deleting cases I may have a dataset with 7 million+ rows and 50+ columns. Will R handle a datafile of this size? If so, how? Thank you in advance, Tom Volscho

Re: [R] Enormous Datasets

2004-11-18 Thread Peter Dalgaard
Thomas W Volscho [EMAIL PROTECTED] writes: Dear List, I have some projects where I use enormous datasets. For instance, the 5% PUMS microdata from the Census Bureau. After deleting cases I may have a dataset with 7 million+ rows and 50+ columns. Will R handle a datafile of this size? If so,

RE: [R] Enormous Datasets

2004-11-18 Thread Vadim Ogranovich
PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Thomas W Volscho Sent: Thursday, November 18, 2004 12:11 PM To: [EMAIL PROTECTED] Subject: [R] Enormous Datasets Dear List, I have some projects where I use enormous datasets. For instance, the 5% PUMS microdata from the Census Bureau

RE: [R] Enormous Datasets

2004-11-18 Thread Liaw, Andy
It depends on what you want to do with that data in R. If you want to play with the whole data, just storing it in R will require more than 2.6GB of memory (assuming all data are numeric and are stored as doubles): 7e6 * 50 * 8 / 1024^2 [1] 2670.288 That's not impossible, but you'll need to be