Hi all, many further thanks for the help I recently received on setting 
up data for repeated measures analysis, thanks to the help I got here I 
have been able to do that analysis. 

A further question:  it seems one could come up with justifications for 
several ways to compute outliers in a dataset. For example with the DV 
"reaction time," one could exclude all measurements > 2 STDs away from 
mean.  But, the overall mean of all measures? or just for each level of 
one of the factors? Or considering both factors, as in, calculate 
outliers for each cell?

In the latter case, one would be decreasing within-cell variance 
relative to overall variance, and thus, increasing the F statistic 
calculated between cells. ...right?  

But if one calculates the outliers from the entire dataset, one 
decreases overall variance relative to between-group or between-cell 
variance, thus artificially deflating the F value.

What approach should one take? Reaction time outliers are most often due 
to distraction or a missed cue, so that it is usually not necessary to 
individually justify excluding each single outlier, and a way to "en 
masse" delete the offenders without harming the statistical model is 
sought.

Many Thanks
Jim

-- 
WATCH OUT for spam block

To reply by email, remove the spam block


=================================================================
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
                  http://jse.stat.ncsu.edu/
=================================================================

Reply via email to