You will need to be more specific (and careful). You talk about large datafiles increase the size of records
so is the problem 1) The size of the files 2) The size of the records 3) Storage for the R objects you are creating, in some unstated way 4) Something else ? Without more details, the most helpful advice we can give is A) Use a 64-bit Linux/Unix OS (e.g. AMD64/Opteron) B) Add as much RAM as you can afford/fit. On Mon, 4 Oct 2004, Greg Butler wrote: > Hi, > > I have been enjoying r for some time now, but was wondering about working > with larger data files. When I try to load in big files with more than > 20,000 records, the programs seems unbable to store all the records. Is > there some way that I can increase the size of records that I work with? > Ideally I would like to work with census data which can hold a million > records. > PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html Please do, and tell us the OS you are using and resources you have (e.g. RAM). -- Brian D. Ripley, [EMAIL PROTECTED] Professor of Applied Statistics, http://www.stats.ox.ac.uk/~ripley/ University of Oxford, Tel: +44 1865 272861 (self) 1 South Parks Road, +44 1865 272866 (PA) Oxford OX1 3TG, UK Fax: +44 1865 272595 ______________________________________________ [EMAIL PROTECTED] mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html