My choices are (in the order of my preference): - use connections and readLines()/strsplit()/etc. in R to process the file a chunk at a time
- use cut/paste/grep/etc., perhaps within pipe() in R - use awk, perhaps within pipe() in R - Python is my last resort, as I'm not familiar with it The first preference is to do it all in R, mostly for the reason that I can keep track of what was done all in one place (the R script or function). Andy > From: David Mitchell > > Hello list, > > I'm regularly in the position where I have to do a lot of data > manipulation, in order to get the data I have into a format R is happy > with. This manipulation would generally be in one of two forms: > - getting data from e.g. text log files into a tabular format > - extracting sensible sample data from a very large data set (i.e. too > large for R to handle) > > In general, I use Perl or Python to do the task; I'm curious as to > what others use when they hit the same problem. > > Regards > > Dave Mitchell > > ______________________________________________ > [EMAIL PROTECTED] mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide! > http://www.R-project.org/posting-guide.html > > ______________________________________________ [EMAIL PROTECTED] mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
