Re: [R] R-help Digest, Vol 37, Issue 12
Hi r-users, I would like to know if R have any solution to the Address standardization. The problem is to classify a database of addresses with the real addresses of a streets of Spain. Ideally, I would like to assign Postal code, census data and other geographic information. If this is not possible I would like to know solutions in R about text mining, text classification, distance within text data,... Any help will be appreciate Thanks in advance Ferran Carrascosa __ R-help@stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
[R] Weighting factor
Hi everyone, I would like some package of R or any help to solve the next problem with a weighting fatcors: Giving a data matrix with dichotomous (2 or more) variables in columns and individuals in rows, and also a theorical distribution of the dichotomous variables I would like a row weight (one per individual) that transform the real distribution of variables to the theorical distribution. Thanks in advance and best regards, -- Ferran Carrascosa __ R-help@stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
Re: [R] Weighting factor
Uau! This package is very interesting, very thank you Thomas. Thanks to all! 2005/9/27, Thomas Lumley [EMAIL PROTECTED]: On Tue, 27 Sep 2005, Ferran Carrascosa wrote: Hi everyone, I would like some package of R or any help to solve the next problem with a weighting fatcors: Giving a data matrix with dichotomous (2 or more) variables in columns and individuals in rows, and also a theorical distribution of the dichotomous variables I would like a row weight (one per individual) that transform the real distribution of variables to the theorical distribution. You should be able to do this with postStratify() in the survey package (if you have joint distributions for the variables) or calibrate() if you have only marginal distributions. You would need to create a survey design object, use postStratify() or calibrate() on it, then use weights() to extract the weights. -thomas -- Ferran Carrascosa __ R-help@stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
Re: [R] memory
Thanks Prof Brian for your answers, I have read about 'ref' package to work with more efficient memory work. Anybody know if this package could help me to work with a 700.000 x 10.000 matrix? I will have problems with ref package on: - Limit of 2 Gb in R for Windows. -The maximum cells in one object 2*10^9 (aprox.) Thanks in advance, -- Ferran Carrascosa 2005/8/30, Prof Brian Ripley [EMAIL PROTECTED]: On Mon, 29 Aug 2005, Ferran Carrascosa wrote: Hi, I have a matrix with 700.000 x 10.000 cells with floating point data. I would like to work with the entire table but I have a lot of memory problems. I have read the ?memory I work with Win 2000 with R2.1.0 The only solution that I have applied is: memory.limit(size=2048) But now my problems are: - I need to work with more than 2 Gb. How I can exceed this limit? Re-read the rw-FAQ, or (preferably) get a more capable OS on a 64-bit CPU. - When apply some algorithms, the maximum cells in one object 2*10^9 (aprox.) is reached. You will never get that many cells (that is the address space in bytes, and they are several bytes each). Please do as the posting guide asks and report accurately what happened. Please could you send me some advises/strategies about the work with large amount of data in R? R have a way to work with less memory needs? Your matrix has 7e09 cells (assuming you are using . as a thousands separator) and needs 5.6e10 bytes to store. Your OS has a memory address limit of 3.2e09 bytes. Don't blame R for being limited by your OS. -- Brian D. Ripley, [EMAIL PROTECTED] Professor of Applied Statistics, http://www.stats.ox.ac.uk/~ripley/ University of Oxford, Tel: +44 1865 272861 (self) 1 South Parks Road, +44 1865 272866 (PA) Oxford OX1 3TG, UKFax: +44 1865 272595 __ R-help@stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html
[R] memory
Hi, I have a matrix with 700.000 x 10.000 cells with floating point data. I would like to work with the entire table but I have a lot of memory problems. I have read the ?memory I work with Win 2000 with R2.1.0 The only solution that I have applied is: memory.limit(size=2048) But now my problems are: - I need to work with more than 2 Gb. How I can exceed this limit? - When apply some algorithms, the maximum cells in one object 2*10^9 (aprox.) is reached. Please could you send me some advises/strategies about the work with large amount of data in R? R have a way to work with less memory needs? Thanks in advance, -- Ferran Carrascosa __ R-help@stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html