James
 
The main reason for the adjusted R^2  (Fisher) is that it is less biased than 
the ordinary R^2.  The ordinary R^2 has a positive bias that is a function of 
the true Rho^2, the number of predictors p, and the sample size n.  The maximum 
bias occurs at Rho^2 = 0, where the expected R^2 is p/(n-1). The adjusted R^2 
has a slightly negative bias (max being on the order of -1/2n at Rho^2 = .5) 
which is not a function of p.  
 
In your example, the R^2 for the 1st equation will be 1 even if Rho^2 = 0, but 
the expected R^2 will be Rho^2 + .04 for the second.  ( I am interpreting 
"parameters" as "predictors", which is strictly speaking not true, as the 
regression intercept and error variance are also parameters.)  The adjR^2 will 
have max expected bias of -.1 in the first and -0.005 in the second.
 
Any between-regression comparions using R^2 will founder on differences in bias 
induced by differences in Rho^2, p, and n. Any between-regressions comparisons 
using adjR^2 will be founder on differences in bias induced by differences in 
Rho^2 and n. However, the maximum possible difference in bias for adjR^2 may 
not be large. 
 
Note also 

1.      
        The standard errors of the estimators should also be taken in account 
in such comparisons.
2.      
        There is an unbiased estimator of Rho^2 (Olkin and Pratt); and
3.      
        There is another adjR^2 has slightly better MSE than the Fisher adjR^2.
4.      
        There is a difference in results if the predictors are considered fixed 
rather than multivariate normal (Barten)

Joe
 
References
 
Barten AP. Note on unbiased estimation of the squared multiple correlation 
coefficient. Statistica Neerlandica, 1962, 16, 151-163
 
Fisher RA. The influence of rainfall in the yield of wheat at Rothamstead.  
Philosophical Transactions of the Royal Society of London, Series B, 1924, 213, 
89-142.
 
Lucke JF and Embreston SE. Biases and mean squared errors of estimators of 
multinormal squared multiple correlation. Journal of Education Statistics, 
1984, 9(3), 183-192.
 
Olkin I and Pratt JW. Unbiased estimation of certain correlation coefficients.  
Annals of Mathematical Statistics, 1958, 29, 201-211.

________________________________

From: [EMAIL PROTECTED] on behalf of James Salsman
Sent: Fri 6/17/2005 4:16 PM
To: r-help@stat.math.ethz.ch
Subject: [R] adjusted R^2 vs. ordinary R^2



I thought the point of adjusting the R^2 for degrees of
freedom is to allow comparisons about goodness of fit between
similar models with different numbers of data points.  Someone
has suggested to me off-list that this might not be the case.

Is an ADJUSTED R^2 for a four-parameter, five-point model
reliably comparable to the adjusted R^2 of a four-parameter,
100-point model?  If such values can't be reliably compared
with one another, then what is the reasoning behind adjusting
R^2 for degrees of freedom?

What are the good published authorities on this topic?

Sincerely,
James Salsman

______________________________________________
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html



        [[alternative HTML version deleted]]

______________________________________________
R-help@stat.math.ethz.ch mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide! http://www.R-project.org/posting-guide.html

Reply via email to