Dear List,
I have some projects where I use enormous datasets. For instance, the 5% PUMS
microdata from the Census Bureau. After deleting cases I may have a dataset
with 7 million+ rows and 50+ columns. Will R handle a datafile of this size?
If so, how?
Thank you in advance,
Tom Volscho
Thomas W Volscho [EMAIL PROTECTED] writes:
Dear List, I have some projects where I use enormous datasets. For
instance, the 5% PUMS microdata from the Census Bureau. After
deleting cases I may have a dataset with 7 million+ rows and 50+
columns. Will R handle a datafile of this size? If so,
PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Thomas
W Volscho
Sent: Thursday, November 18, 2004 12:11 PM
To: [EMAIL PROTECTED]
Subject: [R] Enormous Datasets
Dear List,
I have some projects where I use enormous datasets. For
instance, the 5% PUMS microdata from the Census Bureau
It depends on what you want to do with that data in R. If you want to play
with the whole data, just storing it in R will require more than 2.6GB of
memory (assuming all data are numeric and are stored as doubles):
7e6 * 50 * 8 / 1024^2
[1] 2670.288
That's not impossible, but you'll need to be