On Tue, 30 May 2000, Karen Scheltema wrote:
> Can someone enlighten me about how partial least squares regression
> works to handle multicollinearity.
Depends partly on whether you're looking at real, or spurious,
multicollinearity; and may depend on where the multicollinearity
actually arises. We may need more description of your particular
problem for a useful conversation.
What I understand by "partial regression" is not really different from
multiple regression: it involves "partialling" out the effects of
various predictors both from other (in general susbequent) predictors
and from the response variable, often either in a sequential manner or
in a way that implies a sequence by assigning an order of hierarchical
importance, if you will, to the several predictors.
If the multicollinearity arises from artificial variables computed as
the product of two or more raw variables, usually in seeking evidence
for or against the presence of interactions between those raw variables,
there is an obvious sort of hierarchy:
raw variables > 2-way interactions > 3-way interactions > ...
If the multicollinearity is "built in", so to speak, because there really
exists a (near-)linear combination among the predictors (or even more
than one such combination), one may need to decide which variable(s) to
EXclude in order to avoid the multicollinearity: this is another way of
saying that one needs to assign a hierarchy of importance to the
predictors.
In any event, the main problem with multicollinearity is
computational: in finite precision, estimation becomes unreliable as the
apparent zero-order correlations become less distinguishable from 1.000...
The most effective approach to the problem that I know of is to (begin
to) orthogonalize at least some of the predictors with respect to some
or all of the others. For an example applying this idea in practice
(where multicollinearity arose from computing raw interaction variables)
see my paper on the Minitab web site (www.minitab.com and look for
Resources, then White Papers).
> Can SPSS do partial least squares regression?
Yes, if we're talking on the same wavelength. Requires computing
residual variables and adjoining them to the variables in the data set,
then using them as predictors in place of the products (or raw variables)
they're residuals from.
-- DFB.
------------------------------------------------------------------------
Donald F. Burrill [EMAIL PROTECTED]
348 Hyde Hall, Plymouth State College, [EMAIL PROTECTED]
MSC #29, Plymouth, NH 03264 603-535-2597
184 Nashua Road, Bedford, NH 03110 603-471-7128
===========================================================================
This list is open to everyone. Occasionally, less thoughtful
people send inappropriate messages. Please DO NOT COMPLAIN TO
THE POSTMASTER about these messages because the postmaster has no
way of controlling them, and excessive complaints will result in
termination of the list.
For information about this list, including information about the
problem of inappropriate messages and information about how to
unsubscribe, please see the web page at
http://jse.stat.ncsu.edu/
===========================================================================