Re: [R] R-help Digest, Vol 37, Issue 12

2006-03-12 Thread Ferran Carrascosa
Hi r-users,

I would like to know if R have any solution to the Address standardization.
The problem is to classify a database of addresses with the real
addresses of a streets of Spain. Ideally, I would like to assign
Postal code, census data and other geographic information.

If this is not possible I would like to know solutions in R about text
mining, text classification, distance within text data,...

Any help will be appreciate

Thanks in advance

Ferran Carrascosa

__
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html


[R] Weighting factor

2005-09-27 Thread Ferran Carrascosa
Hi everyone,

I would like some package of R or any help to solve the next problem
with a weighting fatcors:

Giving a data matrix with dichotomous (2 or more) variables in columns
and individuals in rows, and also a theorical distribution of the
dichotomous variables I would like a row weight (one per individual)
that transform the real distribution of variables to the theorical
distribution.

Thanks in advance and best regards,

--
Ferran Carrascosa

__
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html


Re: [R] Weighting factor

2005-09-27 Thread Ferran Carrascosa
Uau!
This package is very interesting, very thank you Thomas.

Thanks to all!

2005/9/27, Thomas Lumley [EMAIL PROTECTED]:
 On Tue, 27 Sep 2005, Ferran Carrascosa wrote:

  Hi everyone,
 
  I would like some package of R or any help to solve the next problem
  with a weighting fatcors:
 
  Giving a data matrix with dichotomous (2 or more) variables in columns
  and individuals in rows, and also a theorical distribution of the
  dichotomous variables I would like a row weight (one per individual)
  that transform the real distribution of variables to the theorical
  distribution.
 

 You should be able to do this with postStratify() in the survey package
 (if you have joint distributions for the variables) or calibrate() if you
 have only marginal distributions.

 You would need to create a survey design object, use postStratify() or
 calibrate() on it, then use weights() to extract the weights.

-thomas



--
Ferran Carrascosa

__
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html


Re: [R] memory

2005-08-30 Thread Ferran Carrascosa
Thanks Prof Brian for your answers,
I have read about 'ref' package to work with more efficient memory
work. Anybody know if this package could help me to work with a
700.000 x 10.000 matrix?

I will have problems with ref package on:
- Limit of 2 Gb in R for Windows.
-The maximum cells in one object 2*10^9 (aprox.)

Thanks in advance,
-- 
Ferran Carrascosa


2005/8/30, Prof Brian Ripley [EMAIL PROTECTED]:
 On Mon, 29 Aug 2005, Ferran Carrascosa wrote:
 
  Hi,
 
  I have a matrix with 700.000 x 10.000 cells with floating point data.
  I would like to work with the entire table but I have a lot of memory
  problems. I have read the ?memory
  I work with Win 2000 with R2.1.0
 
  The only solution that I have applied is:
  memory.limit(size=2048)
 
  But now my problems are:
  - I need to work with more than 2 Gb. How I can exceed this limit?
 
 Re-read the rw-FAQ, or (preferably) get a more capable OS on a 64-bit CPU.
 
  - When apply some algorithms, the maximum cells in one object 2*10^9
  (aprox.) is reached.
 
 You will never get that many cells (that is the address space in bytes,
 and they are several bytes each).  Please do as the posting guide asks
 and report accurately what happened.
 
  Please could you send me some advises/strategies about the work with
  large amount of data in R?
 
  R have a way to work with less memory needs?
 
 Your matrix has 7e09 cells (assuming you are using . as a thousands
 separator) and needs 5.6e10 bytes to store.  Your OS has a memory address
 limit of 3.2e09 bytes.  Don't blame R for being limited by your OS.
 
 --
 Brian D. Ripley,  [EMAIL PROTECTED]
 Professor of Applied Statistics,  http://www.stats.ox.ac.uk/~ripley/
 University of Oxford, Tel:  +44 1865 272861 (self)
 1 South Parks Road, +44 1865 272866 (PA)
 Oxford OX1 3TG, UKFax:  +44 1865 272595


__
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html


[R] memory

2005-08-29 Thread Ferran Carrascosa
Hi,

I have a matrix with 700.000 x 10.000 cells with floating point data.
I would like to work with the entire table but I have a lot of memory
problems. I have read the ?memory
I work with Win 2000 with R2.1.0

The only solution that I have applied is:
 memory.limit(size=2048)

But now my problems are:
- I need to work with more than 2 Gb. How I can exceed this limit?
- When apply some algorithms, the maximum cells in one object 2*10^9
(aprox.) is reached.

Please could you send me some advises/strategies about the work with
large amount of data in R?

R have a way to work with less memory needs?

Thanks in advance,
-- 
Ferran Carrascosa

__
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html