Re: Applied analysis question

2002-02-28 Thread Dennis Roberts

At 07:37 AM 2/28/02 -0800, Brad Anderson wrote:

I think a lot of folks just run standard analyses or arbitrarily apply
some normalizing transformation because that's whats done in their
field.  Then report the results without really examining the
underlying distributions.  I'm curious how folks procede when they
encounter very goofy distrubions.  Thanks for your comments.

i think the lesson to be gained from this is that, we seem to be focusing 
on (or the message that students and others get) getting the analysis DONE 
and summarizied ... and with most standard packages ... that is relatively 
easy to do

for example, you talk about a simple regression analysis and then show them 
in minitab that you can do that like: mtb regr 'height' 1 'weight' and, 
when they do it, lots of output comes out BUT, the first thing is the best 
fitting straight line equation like:

The regression equation is
Weight = - 205 + 5.09 Height

and THAT's where they start AND stop (more or less)

while software makes it rather easy to do lots of prelim inspection of 
data, it also makes it very easy to SKIP all that too

before we do any serious analysis ... we need to LOOK at the data ... 
carefully ... make some scatterplots (to check for outliers, etc.), to look 
at some frequency distributions ON the variables, to even just look at the 
means and sds ... to see if some serious restriction of range issue pops up 
...

THEN and ONLY then, after we get a feel for what we have ... THEN and ONLY 
then should we be doing the main part of our analysis ... ie, testing some 
hypothesis or notion WITH the data (actually, i might call the prelims the 
MAIN part but, others might disagree)

we put the cart before the horse ... in fact, we don't even pay any 
attention to the horse

unfortunately, far too much of this is caused by the dominant and 
preoccupation of doing significance tests so we run routines that give us 
these p values and are done with it ... without  paying ANY attention to 
just looking at the data

my 2 cents worth



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Means of semantic differential scales

2002-02-28 Thread Dennis Roberts

At 09:51 AM 2/28/02 -0800, Jay Tanzman wrote:


I partially did this, insofar as I ran Pearson and Spearman correlations 
between
several of the scales and, not surprisingly, the two correlation coefficients
and their p-values were similar.  that issue is entirely a separate 
one since the rank order FORMULA was derived from the pearson ...




  Dr. Kim was not impressed.

-Jay

i hate to ask this question but, what the heck, spring break is near so i will

if your boss, dr. kim??? ... seems so knowledgeable about what the data are 
and what is and is not appropriate to do with the data, why is not dr. kim 
doing the analysis?

this reminds me of assigning a task to someone and, doing so much 
micro-managing that ... it would have been better off doing it oneself ...



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Applied analysis question

2002-02-27 Thread Dennis Roberts

At 04:11 PM 2/27/02 -0500, Rich Ulrich wrote:

Categorizing the values into a few categories labeled,
none, almost none,   is one way to convert your scores.
If those labels do make sense.

well, if 750 has the same numerical sort of meaning as 0 (unit wise) ... in 
terms of what is being measured then i would personally not think so SINCE, 
the categories above 0 will encompass very wide ranges of possible values

if the scale was # of emails you look at in a day ... and 1/3 said none or 
0 ... we could rename the scale 0 = not any, 1 to 50 as = some, and 51 to 
750 as = many (and recode as 1, 2, and 3) .. i don't think anyone who just 
saw the labels ... and were then asked to give some extemporaneous 'values' 
for each of the categories ... would have any clue what to put in for the 
some and many categories ... but i would predict they would seriously 
UNderestimate the values compared to the ACTUAL responses

this just highlights that for some scales, we have almost no 
differentiation at one end where they pile up ... perhaps (not saying one 
could have in this case) we could have anticipated this ahead of time and 
put scale categories that might have anticipated that

after the fact, we are more or less dead ducks

i would say this though ... treating the data only in terms of ranks ... 
does not really solve anything ... and clearly represents being able to say 
LESS about your data or interrelationships (even if the rank order r is .3 
compared to the regular pearson of about 0) ... than if you did not resort 
to only thinking about the data in rank terms




--
Rich Ulrich, [EMAIL PROTECTED]
http://www.pitt.edu/~wpilib/index.html


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Means of semantic differential scales

2002-02-27 Thread Dennis Roberts

At 01:39 PM 2/27/02 -0600, Jay Warner wrote:

  
  Not stressful 1__ 2__ 3__ 4__ 5__ 6__ 7__ Very stressful

just out of curiosity ... how many consider the above to be an example of a 
bipolar scale?

i don't

now, if we had an item like:

sad  happy
1  . 7

THEN the mid point becomes much more problematic ...

since being a 4 ... is neither a downer nor upper

now, a quick search found info from ncs about the 16pf personality scale 
... it shows 16 BIpolar dimensions as:

Bipolar Dimensions of Personality
Factor A Warmth (Cool vs Warm)
Factor B Intelligence (Concrete Thinking vs Abstract Thinking)
Factor C Emotional Stability (Easily Upset vs Calm)
Factor E Dominance (Not Assertive vs Dominant)
Factor F Impulsiveness (Sober vs Enthusiastic)
Factor G Conformity (Expedient vs Conscientious)
Factor H Boldness (Shy vs Venturesome)
Factor I Sensitivity (Tough-Minded vs Sensitive)
Factor L Suspiciousness (Trusting vs Suspicious)
Factor M Imagination (Practical vs Imaginative)
Factor N Shrewdness (Forthright vs Shrewd)
Factor O Insecurity (Self-Assured vs Self-Doubting)
Factor Q1 Radicalism (Conservative vs Experimenting)
Factor Q2 Self-Sufficiency (Group-Oriented vs Self-Sufficient)
Factor Q3 Self-Discipline (Undisciplined vs Self-Disciplined)
Factor Q4 Tension (Relaxed vs Tense)

let's take the one ... shy versus venturesome ...

now, we could make a venturesome scale by itself ...

0 venturesomeness .. (up to)  very venturesome 7

does 0 = shy seems like if the answer is no ... then we might have a 
bipolar scale ... if the answer is yes ... then we don't



  It could be the use of the particular bipolars
  not stressful and very stressful.
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: What is an outlier ?

2002-02-25 Thread Dennis Roberts

of course, if one has control over the data, checking the coding and making 
sure it is correct is a good thing to do

if you do not have control over that, then there may be very little you can 
do with it and in fact, you may be totally UNaware of an outlier problem

i see as a potentially MUCH larger problem when ONLY certain summary 
statistics are shown without any basic tallies/graphs displayed so, IF 
there are some really strange outlier values, it usually will go undetected ...

correlations are ONE good case in point ... have a look at the following 
scatterplot ... height in inches and weight in pounds ... from the pulse 
data set in minitab


  -  *
  -
   300+
  -
  Weight  -
  - 2
  - 2  224 32
   150+   ** 3458*454322*
  -*53*3*535  2
  -  **
--+-+-+-+-+-+Height
   32.0  40.0  48.0  56.0  64.0  72.0

now, the actual r between the X and Y is -.075 ... and of course, this 
seems strange but, IF you had only seen this in a matrix of r values ... 
you might say that perhaps there was serious range restriction that more or 
less wiped out the r in this case ...  but even the desc. stats might not 
adequately tell you of this problem

IF you had the scatterplot, you probably would figure out REAL quick that 
there is a PROBLEM with one of the data points ...

in fact, without that one weird data point, the r is about .8 ... which 
makes a lot better sense when correlating heights and weights of college 
students


At 09:06 PM 2/25/02 +, Art Kendall wrote:

--6F47CB3D3B10A10A3E9B064C
Content-Type: text/plain; charset=us-ascii
Content-Transfer-Encoding: 7bit

An outlier is any value for a variable that is suspect given the
measurement system, common sense,  other values for the variable in
the data set, or  the values a case has on other variables.
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



RE: Chi-square chart in Excel

2002-02-21 Thread Dennis Roberts

sure is easy in minitab ... one can draw a very nice curve (it's easy but, 
hard to post here) but, to make a distribution easy for viewing we can

MTB  rand 10 c1;  generated 10 values from
SUBC chis 4.   a chi square distribution with 4 degrees of freedom
MTB  dotp c1

Dotplot: C1


Each dot represents up to 778 points
 .
.::.
:
   .::.
   :
   ::.
   :.
  ::. ..
  +-+-+-+-+-+---C1
0.0   6.0  12.0  18.0  24.0  30.0

MTB  desc c1

Descriptive Statistics: C1


Variable N   Mean Median TrMean  StDevSE Mean
C1  10 4.0123 3.3729 3.7727 2.8350 0.0090

Variable   MinimumMaximum Q1 Q3
C1  0.008029.4143 1.9236 5.4110

not quite as fancy as the professional graph but, will do in a pinch



At 06:27 PM 2/21/02 -0800, David Heiser wrote:


-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED]]On Behalf Of Ronny Richardson
Sent: Wednesday, February 20, 2002 7:29 PM
To: [EMAIL PROTECTED]
Subject: Chi-square chart in Excel


Can anyone tell me how to produce a chart of the chi-square distribution in
Excel? (I know how to find chi-square values but not how to turn those into
a chart of the chi-square curve.)


Ronny Richardson
---
Excel does not have a function that gives the Chi-Square density

The following might be helpful regarding future graphs. It is a fraction of
a larger package I am preparing. It is awkward to present it in .txt
format.


DISTRIBUTIONDENSITY CUMMULATIVE I   NVERSE
BetaBETADIST 
BETAINV
BinomialBINOMST CRITBINOM
Chi-Square  CHIDIST CHINV
Exponential EXPONDIST   EXPONDIST
F   FDIST   FINV
Gamma   GAMMADIST   GAMMADIST 
GAMMAINV
Hyper geometric HYPGEOMDIST
Log Normal  LOGNORMDIST LOGINV
Negative Binomial   NEGBINOMDIST
Normal(with parameters) NORMDISTNORMDISTNORMINV
Normal (z 
values)   NORMSDIST   NORMSINV
Poisson POISSON
t   TDIST   TINV
Weibull WEIBULL

You have to build a column (say B) of X values.

Build an expression for column C calculating the Chi-Square density, given
the x value in col B and the df value in A1.

It would be =exp(($A$1/2)*LN(2) + GAMMALN($A$1/2) + (($A$1/2)-1)*LN(B1) -
B1/2) without the quotes.
You can equation-drag this cell down column C for each X value.

Now build a smoothed scatter plot graph as series 1 with the X value column
B and the Y value as column C.

DAHeiser



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Correlations-statistics

2002-02-20 Thread Dennis Roberts

well, one simple way would be to add B and C ... then correlate with A

if these are radically different scales, convert to z scores first



At 02:05 AM 2/20/02 -0800, Holger Boehm wrote:
Hi,

I have calculated correlation coefficients between sets of parameters
(A) and (B) and beween (A) and (C).
Now I would like to determine the correlation between (A) and (B
combined with C). How can I combine the two parameters (B) and (C),
what kind of statistical method has to be applied?

Thanks for your tips,

Holger Boehm


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: What is an experiment ?

2002-02-20 Thread Dennis Roberts

At 03:59 PM 2/20/02 -0300, Voltolini wrote:
Hi,

I was reading a definition of  experiment in science to be used in a
lecture and the use of treatments and controls are an important feature of
an experiment but my doubt is... is it possible to plan an experiment
without a control and call this as an experiment ?

For example, in a polluted river basin there is a gradient of contamination
and someone are interested in to compare the fish diversity in ten rivers of
this basin. Then, the pollution level are the treatment (with ten levels)
but if there is not a clean river in the basin, I cannot use a control !

Is this an experiment anyway ?

the main issue is CONTROL OVER ... that the experimenter exerts over an 
independent variable

if you want to compare diversity of fish ACROSS rivers ... and you find a 
difference, what does this necessarily have to do with contamination?

i see three variables here ... diversity of fish ... rivers ... level of 
contamination (ie, where the gradients are different)

what are you trying to show impacts on what?


Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Statistical Distributions

2002-02-19 Thread Dennis Roberts

At 07:34 AM 2/19/02 -0500, Herman Rubin wrote:

I do not see this.  The binomial distribution is a natural
one; the normal distribution, while it has lots of mathematical
properties, is not.

i don't know of any distribution that is natural ... what does that mean? 
inherent in the universe? all distributions are human made ... in the sense 
that WE observe events ... and, find some function that links events to 
probabilities

all of mathematics ... and statistics too as an offshoot ... is made up




Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Statistical Distributions

2002-02-18 Thread Dennis Roberts

not to disagree with alan but, my goal was to parallel what glass and 
stanley did and that is all ...seems like there are all kinds of 
distributions one might discuss AND, there may be more than one order that 
is acceptable

most books of recent vintage (and g and s was 1970) don't even discuss what 
g and s did

but, just for clarity sake ... are you saying that the nd is a logical 
SECOND step TO the binomial or, that if you look at the binomial, one could 
(in many circumstances of n and p) say that the binomial is essentially a 
nd (very good approximation).. ?

the order i had for the nd, chis square, F and t seemed to make sense but, 
i don't necessarily buy that one NEED to START with the binominal

certainly, however, if one talks about the binomial, then the link to the 
nd is a must

At 06:36 PM 2/17/02 -0500, Timothy W. Victor wrote:
I also think Alan's idea is sound. I start my students off with some
binomial expansion theory.

Alan McLean wrote:
 
  This is a good idea, Dennis. I would like to see the sequence start with
  the binomial - in a very real way, the normal occurs naturally as an
  'approximation' to the binomial.
 

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Statistical Distributions

2002-02-18 Thread Dennis Roberts

addendum

if one manipulates n and p in a binomial and, gets to a point where a 
person would say (or we would say as the instructor) that what you see is 
very similar to ... and might even be approximated well by ... the nd 
...  this MEANS that the nd came first in the sense that one would have to 
be familiar with that before you could draw the parallel



At 06:36 PM 2/17/02 -0500, Timothy W. Victor wrote:
I also think Alan's idea is sound. I start my students off with some
binomial expansion theory.

Alan McLean wrote:
 
  This is a good idea, Dennis. I would like to see the sequence start with
  the binomial - in a very real way, the normal occurs naturally as an
  'approximation' to the binomial.

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Which is faster? ziggurat or Monty Python (or maybe something else?)

2002-02-18 Thread Dennis Roberts

At 09:50 AM 2/18/02 +, Ian Buckner wrote:
We generate pairs of properly distributed Gaussian variables at
down to 10nsec intervals, essential in the application. Speed can
be an issue, particularly in real time situations.

Ian

wow ... how our perspectives have changed! back in grad school, we had some 
cdc mainframe over in the next building and, we would go over and stick our 
stack of cards through the slot .. and then come back the NEXT  DAY ... 
to pick them up ... we just hoped we hadn't put a , or other bad character 
someplace and would have to wait another day




Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Numerical recipes in statistics ???

2002-02-18 Thread Dennis Roberts

what is it you wanted to cook?

At 01:35 PM 2/18/02 -0800, The Truth wrote:
Are there any Numerical Recipes like textbook on statistics and 
probability ?
Just wondering..

Thanks.


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: F-test

2002-02-14 Thread Dennis Roberts

can you bit a bit more specific here? f tests AND t tests are used for a 
variety of things

give us some context and perhaps we can help

at a minimum of course, one is calling for using a test that involves 
looking at the F distribution for critical values ... the other calls for 
using a t distribution but, that still does not really tell you what is 
going on

At 04:41 AM 2/14/02 -0800, Jan wrote:
The question is how do I see the difference when it's asked for an
f-test or a t-test?


Jan


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: If T-Test can not be applied

2002-02-14 Thread Dennis Roberts

it's called the behrens-fisher problem ... there is nothing that says that 
population variances HAVE to be equal

essentially what you do is to be a bit more conservative in your degrees of 
freedom ... most software packages do this as the default ... or at least 
give you the choice between making an assumption of equal population 
variances ... or not

At 11:48 PM 2/14/02 +0100, Matthias wrote:
Hello,

would be nice if someone can give me some advice with regard to the
following problem:

I would like to compare the means of two independent numerical sets of data
whether they are significantly different from each other or not. One of the
two underlying assumption to calculate the T-Test is not given (Variances
are assumed to be NOT equally distributed; but data is normally
distributed). What kind of (non?)parametric-test does exist - instead of the
T-Test - to calculate possible differences in the two means?
I'm using SPSS for further calculations.

Thank you for your time and help,

Matthias





=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: one-way ANOVA question

2002-02-13 Thread Dennis Roberts

At 09:21 AM 2/13/02 -0600, Mike Granaas wrote:
On Fri, 8 Feb 2002, Thomas Souers wrote:
 
  2) Secondly, are contrasts used primarily as planned comparisons? If 
 so, why?
 

I would second those who've already indicated that planned comparisons are
superior in answering theoretical questions and add a couple of comments:

another way to think about this issue is: what IF we never had ... nor will 
in the future ... the overall omnibus F test?

would this help us or hurt us in the exploration of the 
experimental/research questions of primary interest?

i really don't see ANY case that it would hurt us ...

and, i can't really think of cases where doing the overall F test helps us ...

i think mike's point about planning comparisons making us THINK about what 
is important to explore in a given study ... is really important because, 
we have gotten lazy when it comes to this ... we take the easy way out of 
testing all possible paired comparisons when, it MIGHT be that NONE of 
these are really the crucial things to be examined




Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: one-way ANOVA question

2002-02-08 Thread Dennis Roberts

At 10:37 AM 2/8/02 -0800, Thomas Souers wrote:

2) Secondly, are contrasts used primarily as planned comparisons? If so, why?

well, in the typical rather complex study ... all pairs of possible mean 
differences (as one example) are NOT equally important to the testing of 
your theory or notions

so, why not set up ahead of time ... THOSE that are (not necessarily 
restricted to pairs) you then follow ... let the other ones alone

no law says that if you had a 3 by 4 by 3 design, that the 3 * 4 * 3 = 36 
means all need pairs testing ... in fact, come combinations may not even 
make a whole lot of sense EVEN if it is easier to work them into your design


I would very much appreciate it if someone could take the time to explain 
this to me. Many thanks.


Go Get It!
Send FREE Valentine eCards with Lycos Greetings
http://greetings.lycos.com


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: 'Distance' between two normal distributions

2002-02-06 Thread Dennis Roberts

seems, as you have said, depends what you want to do with it

if there is considerable overlap, then whatever distance you use will have 
some of both distributions included ... if there is essentially no overlap 
... then any pair of values ... one from each ...will reflect a real difference

of course, if there is a small difference in means but very large sds ... 
that is one thing wheres ... if there were the same small differences in 
means but, minuscule sds ... that would be another thing

the simple thing would be to use the mean difference but, that really does 
not reflect if there is any overlap between the two and, that seems to be 
part of the issue

At 07:28 PM 2/6/02 +, Francis Dermot Sweeney wrote:
If I have two normal distributions N(m1, s1) and N(m2, s2), what is a
good measure of the distance between them? I was thinking of something
like a K-S distance like max|phi1-phi2|. I know it probably depende on
what I want it for, or what exactly I mean by distance, but any ideas
would be helpful.

Thanks,
Francis.

--

Francis Sweeney
Dept. of Aero/Astro
Stanford U.


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

Dennis Roberts, 208 Cedar Bldg., University Park PA 16802
Emailto: [EMAIL PROTECTED]
WWW: http://roberts.ed.psu.edu/users/droberts/drober~1.htm
AC 8148632401



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: COV

2002-02-02 Thread Dennis Roberts

here is one VERY simple example

the COV is the AVERAGE of the PRODUCTS of the deviations around the means 
... of two variables

if the cov is + .. there is a + relationship between X and Y, if it is -, 
there is a - relationsip between X and Y

X  Y   devX  devY (devX)(devY)
10   20 1-2   - 2
9 24 0 2 0
8 22 -10 0

the sum of the products of the deviations around the two means is -2  
the average is -2 / 3 = -.67 = the covariance

now, some books will divide by n-1 or 2 in this case ... which would = -2/2 
= -1 for the covariance

here is the minitab output ... note, mtb and most software packages will 
divide by n-1


  Y   -  *
  -
  -
  22.5+
  - *
  -
  -
  -
  20.0+   *
  -
+-+-+-+-+-+--X
 8.00  8.40  8.80  9.20  9.60 10.00

MTB  prin c1-c5

Data Display


  Row  X  Y   devX   devY  product

1 10 20  1 -2   -2
2  9 24  0  20
3  8 22 -1  00

MTB  sum c5

Sum of product

Sum of product = -2.
MTB  cova c1 c2

Covariances: X, Y


 XY
X 1.0
Y-1.0  4.0

MTB  NOTE ... THE INTERSECTION OF THE X AND Y OUTPUT ABOVE ... -1 ... IS 
THE COVARIANCE IN THIS PROBLEM

by the way, the CORRELATION is

MTB  corr c1 c2

Correlations: X, Y


Pearson correlation of X and Y = -0.500





At 06:06 PM 2/2/02 +, Maja wrote:
Hi everyone,

we just leanred the cov, and I'm trying to apply the formula
cov(X,Y)=E(XY)-E(X)E(Y) to the following question
f(x,y)=(0.6)^x(0.4)^1-x(0.3)^y(0.52)^1-y(2)^xy where poss. values for X and
Y are x=0,1 and y=0,1.
The prof. never gave us any examples for cov and neither does the text book.
Now I don't know what values to plug in for X and what values to plug in for
Y.

Could someone PLEASE explain to me how am I supposed to know what vales to
plug into the equation???

TNX
Maja




=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: correlation of dependent variables

2002-01-31 Thread Dennis Roberts
://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Definition of Relationship Between Variables (was Re: Eight

2002-01-30 Thread Dennis Roberts




experimental design, whose basic principles are rather simple, is elegant 
and if applied in good ways, can be very informative as to data, variables 
and their impact, etc.

but, please hold on for a moment

when it comes to humans, we have developed some social policies that say:

1. experimental procedures should not harm Ss ... nor pose any unnecessary risk
2. Ss should be informed about what it is they are being asked to 
participate in as far as experiments are concerned
3. Ss are THE ones to make the decision about participation or not (except 
in cases of minors ... we allocate that decision to guardians/parents)

you aren't suggesting, are you, that for the sake of knowledge, we should 
abandon these principles?

so you see, there are serious restrictions (and valid ones at that) when it 
comes to doing experiments with humans ...

for so much of human behavior and things we would like to explore, we are 
unable to do experiments ... it is that plain and simple

yes, while we may be able to gather data on many of these variables of 
interest in other ways ... we are rarely if EVER in the position of being 
able to say with any causative assurance (since we did not experimentally 
manipulate things) that when we do X, Y happens

208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: area under the curve

2002-01-30 Thread Dennis Roberts

unless you had a table comparable to the z table for area under the normal 
distribution ... for EACH different level of skewness ... an exact answer 
is not possible in a way that would be explainable

here is one example that may help to give some approximate idea as to what 
might happen though


  .
.:::.
   .:.
  .:::.
 .:.
.:::.
   :::
  . .:::. .
  +-+-+-+-+-+---C1

the above is a norm. distribution of z scores ... where 1/2 the data are 
above 0 and 1/2 are below

   :
   :.
  .::.
  .
  :.
  .
  . ...   .
  +-+-+-+-+-+---C2
   -5.0  -2.5   0.0   2.5   5.0   7.5

here is a set of z score data that is radically + skewed ... and even 
though it has 0 as its mean, 50% of the area is NOT above the mean of 0 ... 
note the median is down at about -.3 ... so, there is LESS than 50% above 
the mean of 0 ...

this means that for z scores above 0 ... there is not as much area beyond 
(to the right) ... as you would expect if the distribution had been normal ...

so, we can have some approximate idea of what might happen but the exact 
amount of this clearly depends on how much skew you have

MTB  desc c1 c2

Descriptive Statistics: C1, C2


Variable N   Mean Median TrMean  StDevSE Mean
C1   1 0.0008 0.0185 0.0029 1.0008 0.0100
C2   1 0.-0.3098-0.1117 1. 0.0100

Variable   MinimumMaximum Q1 Q3
C1 -3.9468 3.9996-0.6811 0.6754
C2 -0.9946 8.0984-0.7127 0.3921



At 07:27 AM 1/30/02 -0800, Melady Preece wrote:
A student wants to know how one can calculate the area under the curve for 
skewed distributions.  Can someone give me an answer about when a 
distribution is too skewed to use the z table?

Melady

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: cutting tails of samples

2002-01-30 Thread Dennis Roberts
 this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: transformation of dependent variable in regression

2002-01-16 Thread Dennis Roberts

there is nothing from stopping you (is there?) trying several methods that 
are seen as sensible possibilities ... and seeing what happens?

of course, you might find a transformational algorithm that works BEST (of 
those you try) with the data you have but ... A) that still might be an 
optimal solution and/or B) it might be best with THIS data set but ... 
for other similar data sets it might not be

i think the first hurdle you have to hop over is ... does it make ANY sense 
WHATSOEVER to take the data you have collected (or received) and change the 
numbers from what they were in the first place? if the answer is YES to 
that then my A and B comments seem to apply but, if the answer is NO ... 
then neither A nor B seem justifiable

with 2 independent and 1 dependent variables... you have possibilities for 
transforming 0 of them ... 1 of them ... 2 of them ... or all of them and, 
these various combinations of what you do might clearly produce varying results




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: random versus fixed factor

2002-01-15 Thread Dennis Roberts

gee  just a short question to answer!

here is one  part of it

say you were interested in whether different teaching methods impacted on 
how well students learned intro statistics ...

now, if we put our minds to it, there probably are 50 or more different 
ways we could teach a course like this but, there is no way you would be 
able to run an experiment trying out all 50

1. you could take several methods AT random (after you list out all 50) ... 
maybe 3 ... and try ... and, if you find some differences amongst these 3 
... then you might be able to generalize back to the entire set of 50 ... 
ie, there are differences amongst the others too

2. perhaps there are only 3 specific methods you have any interest in (out 
of the 50) ... and are not interested in any of the other 47. so, these 
specific 3 you use in your experiment and perhaps you find some 
differences. well, in this case ... you might be able to say that there ARE 
differences in these 3 but, you could not generalize to the larger set of 
the other 47 (since you did not sample representatively from all 50)

#1 is called a random (independent variable) factor
#2 is called a fixed (independent variable) factor

At 02:39 PM 1/15/02 -0800, Elias wrote:
hi
i am a little confused about this  topic

=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: SAT Question Selection

2002-01-14 Thread Dennis Roberts

for the SAT ... which is still paper and pencil ... you will find multiple 
sections ... math and verbal ... as far as i know ... there usually are 3 
of one and 2 of the other ... the one with 3 has A section that is called 
operational ... which does NOT count ... but is used for trialing new 
items ... revised items ... etc.

don't expect them to tell you which one that is however ...

in a sense ... they are making YOU pay for THEIR pilot work ... and, of 
course, if you happen to really get fouled up on the section that is 
operational and does not count ... it could carry over emotionally to 
another section ... and have some (maybe not much) impact on your 
motivation to do well on that next section

unless it has changed ...



At 05:19 PM 1/14/02 -0500, you wrote:
[cc'd to previous poster; please follow up in newsgroup]

L.C. [EMAIL PROTECTED] wrote in sci.stat.edu:
 Back in my day (did we have days back then?) I recall
 talk of test questions on the SAT. That is, these questions
 were not counted; they were being tested for (I presume)
 some sort of statistical validity.
 
 Does anyone have any statistical insight into the SAT question
 selection process. Does anyone have a specific lead? I can
 find virtually nothing.

I remember reading a good book about the inner operation of ETS
(administers the SATs), with some bits about the test questions
you refer to, but I can't quite remember the title. I've searched
the catalog of my old library, and this _may_ be it:

Lemann, Nicholas.
  The big test : the secret history of the American meritocracy
  New York : Farrar, Straus and Giroux, 1999.

--
Stan Brown, Oak Road Systems, Cortland County, New York, USA
   http://oakroadsystems.com/
What in heaven's name brought you to Casablanca?
My health. I came to Casablanca for the waters.
The waters? What waters? We're in the desert.
I was misinformed.


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Proportionate vs. disproportionate

2002-01-11 Thread Dennis Roberts

1. well, one can consider proportionate ... equal change VALUES ... and i 
think that is one legitimate way to view it ... which is how the video guy 
was talking about it ...

2. one could consider proportionate ... equal change from the BASE ... and 
i think that is legitimate too ... this is clearly NOT how the video guy 
was referring to it to

sure, if one group goes from 60 to 90 ... this is a change of 30 % points ...
and, if another group goes from 30 to 60 ... this is a change of 30% points 
...
i think it is fair to say that the amount of change % points is the same 
... thus proportional

now, another way we can view the data is to say that in the first group ... 
since the base is 60 ... the change of 30 is a 50% change in % points ... 
compared to the base ... whereas in the second group ... the change from 30 
to 60 represents a 100% change from the base ...

now, if the base N is the same ... say 600 people ... 30% of 600 = 180 
people ... no matter if a group change from 60 to 90 OR 30 to 60 ... thus, 
if the BASE n is the same ... then both value of % change AND volume of n 
... mean the same thing

if one group's n = 600 and another group's n = 100 ... these are not the same

but in any case ... and the way we usually look at these poll %ages ... is 
in terms of the absolute value of the % values ... so, in THAT context and 
the way the public usually views these things ... scenario #1 above is how 
the video person presented the data ... and in that context i think his 
presentation did NOT try nor did snow the video viewers

i don't think there is any natural law that says that proportionate or 
disproportionate has to be interpreted in terms of scenario #2 above ...

finally ... i think we are making a mountain out of a molehill in this ... 
to me ... the most important fact from the video was that (regardless of 
change and how you define it) ... whites approved of the president to a FAR 
greater extent than blacks ... and, the second most important fact was 
that AFTER the event ... the approval ratings for BOTH groups went un 
dramatically

if the video guy had made the distinctions in scenarios 1 and 2 above ... 
and had then interpreted the data under both cases ... i think this would 
have helped NONE in conveying to the public the information that i (IMHO) 
think was most important ...

we seem to be trying to find something that the fellow was hiding FROM the 
public when, i don't really think he was trying (nor gallup) to hide 
anything ... he was presenting some data results ... giving one 
interpretation of the results ... and, if WE want to interpret them 
differently ... we can

is that not true for any set of results?




At 10:58 AM 1/11/02 -0500, you wrote:


On Fri, 11 Jan 2002, Dennis Roberts wrote:

  if the polls used similar ns in the samples ... i disagree
 
  now, if the white sample was say 600 and the black sample was 100 ... i
  MIGHT be more likely to agree with the comment below

consider white goes 10% to 15% up 50%, 5%pts
  black goes 66.7% to 100%  up 50%, 33.3%pts

These are proportionate but hardly equivalent

white goes 0 to 50% up infinite %, 50%pts
black goes 50% to 100%  up 100%50pts
same %pts but black is more striking
I can't see any kind of equivalence in either case

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Proportionate vs. disproportionate

2002-01-10 Thread Dennis Roberts

there are two sets of data ... one for georgeDUBU ... and the elder george bush

here is what i glean from the charts

for george w ... the EVENT was sept 11 ... for the elder george bush ... 
the EVENT was the gulf war ... and both were before and after ratings

1. whites approval rating for BOTH ... was much higher than blacks
2. both whites and blacks jumped rather dramatically (in the 30 percent 
range) on AFTER compared to BEFORE
3. to me, proportionate would be both increasing the same approximate % 
... disproportionate would imply large differentials in % changes ...

in neither case were the % jumps the same ... for each bush ... before and 
after ... comparing whites and blacks (assuming the data reported in the 
video is correct) ... so TECHNICALLY ... it is disproportionate ... but ... 
what about approximately ??? i think it is a matter of practical 
differences and semantics ... not really statistically significant 
differences ... given the ns ... it is possible that the difference in THE 
differences MIGHT have been significant ...

here are the values

george w ...

WHITES
pre post change

60   9030

BLACKS

33   6835


elder george

WHITES

64   9026

BLACKS

33   7037

difference between w/b for geore w = 31 versus 35

difference between w/b for elder george = 26 versus 37

now, i would be willing to say that there is less difference in change for 
george w than the elder george ...

in viewing the video ... i did not see that the person really said anything 
categorical about this ... he used the term roughly ... just depends if 
the VIEWER of the video and data wants to think that 4% verus 11% means 
roughly the same change ... or not

thus, i don't think the moderator said anything really wrong ...




At 04:27 PM 1/10/02 +, EugeneGall wrote:
The Gallup organization posted a video to explain why the the increase in
black's job approval for Bush is 'proportionate' to the increase among 
whites.
Both increased by about 30% (60 to 90 for whites, mid thirties to roughly 70%
for blacks), so the increase is proportionate, not disproportionate, since 
both
increases were about 30%.  Unless I'm missing something, and I don't think I
am, this proportionate - disproportionate error is repeated and emphasized
several times in the video.

http://www.gallup.com/poll/Multimedia/video/archived/2002/01/vr020108b.ram

Gene Gallagher
UMASS/Boston


=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Excel vs Quattro Pro

2002-01-09 Thread Dennis Roberts

this is about the most irrelevant argument i have heard ... as though the 
only stat package is SAS ...

there are many excellent stat packages ... even their student trimmed 
down versions are better that excel add ons ...

and, hundreds of institutions have cheap software purchase options ...

at penn state for example ... the full package of minitab is about 90 bucks 
... that's not bad for an excellent tool that will serve one's analysis 
needs well

in addition, students could go to http://www.e-academy.com ... and find 
that they could lease minitab for 6 months for 26 bucks ... or a year for 
50 bucks ...

i challenge any person to try a real package (doesn't have to be minitab) 
and see what you can do and THEN gravitate back to excel's add ons ...

finally, i find the implied notion below that what we need are free 
things ... and that's the way to go ... as the way to operate ... to be 
professionally appalling ...

most institutions SHOULD have a good statistical package on their lab 
systems ... so, students can learn with a good tool

then, when and if they decide that they would like that tool (or another) 
in their professional array of tools ... THEN they could shop around and 
look for some stat package that is within their own or their employer's reach

the bottom line here seems to be:

since excel is free ... and around ... use it. even though we know that it 
was never designed to be a good full serviced package in statistics and 
graphics

we do our students a huge DISservice when we knowingly push tools (just 
because they are cheap or free) RATHER than introduce them to better more 
useful resources

serious companies and institutions and agencies ... DON'T use excel to do 
their mainline statistical work ...

At 03:58 AM 1/9/02 +, you wrote:
Why bother teaching students SAS if nobody can afford their annual license 
fee?
  Spreadsheets works because many people owns MS Office and chances of their
using skills learned in class is greater.

Ken




=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Excel vs. Specialized stats packages (was: Excel vs Quattro Pro)

2002-01-08 Thread Dennis Roberts


rse?

This is an interesting discussion, but the line between a spreadsheet and
stats package is not so clear-cut these days. If you look at how the major
stats packages have developed over the last decade, you can see how they
have copied more and more features from Excel. In fact almost all stats
packages now boast of containing a fully featured built-in spreadsheet for
data entry.

certainly minitab makes no such claim ... their worksheet is NOT a spreadsheet




Looking at the situation from another angle, why can't a spreadsheet be used
for statistical analysis? Granted, some of Excel's built-in statistical
functions leave a lot to be desired and should be used with care. But the
Excel spreadsheet package is still head-and-shoulders above any other
similar product in terms of ease of use, data entry and collection,
presentation, programming interfaces, and it's excellent integration with
the other Office applications.

so, i am not sure this has anything to do with statistical analysis




So if the basic spreadsheet component is sound, and almost all computer and
non-computer literate users can use Excel without problems, why not just
extend Excel's statistical capabilities with reliable accurate statistical
add-ons? Many exist, and we develop a product called Analyse-it for this
very purpose.

i have looked at analyse-it and one other plug in (plus what comes with 
excel) ... and, there just is no comparision between them (well there is 
... and it is not very good) and most of the popular stat packages

A reliable low-cost statistics add-on for Excel can easily bypass these
problems.

unfortunately though, it does not exist

here are the major problems with using excel as a stat package including 
3rd party plugins (off the top of my head)

1. poor data MANAGEMENT capabilities
2. poor and HIGHLY LIMITED graphics
3. highly limited set of routines to select from
4. inability to work with any/many random generation functions (for 
distributions)
5. limited access to important statistical tables

from discussions like this on several lists, it is clear that no argument 
pro or con will sway those who have opted for or agin using excel as the 
statistical analysis tool

but, each side keeps trying

this kind of discussion, though interesting, pales in comparision to a 
discussion we should be having about the over reliance and importance we 
place in statistical analysis in the first place ... and even though i have 
been in this sort of enterprise for more years than you can shake a stick 
at ... the reality is that typical analysis that we do has limited 
practical uses and benefits

the entire area of statistical significance testing is just one case in point



_

James Huntington,
..Analyse-it Software, Ltd.
.
Analyse-it! accurate low-cost statistical software for
Microsoft Excel. For more information  to download a
free evaluation, visit us: http://www.analyse-it.com




=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Measurement Position

2002-01-08 Thread Dennis Roberts

Now that the holidays are behind (most) us, I would like to remind possible 
interested persons in the Asst./Assoc. Professor position opening in 
Measurement at Penn State ... in the Educational Psychology program. We are 
hoping to find someone with SOME experience beyond the doctorate (but, new 
doctoral recipients will be considered) with primary focus instructional 
areas being some combination of IRT, HLM, and SEM

The description of the position can be found at:

http://www.ed.psu.edu/employment/edpsymeas.asp

If you have ANY questions about this position, please contact ME directly at:

mailto:[EMAIL PROTECTED]





_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Correlation problem

2002-01-07 Thread Dennis Roberts

sum of deviations around a mean always = 0


X-X
1-3.5=2.5
2-3.5=-1.5
3-3.5=-0.5
3-3.5=-0.5
4-3.5=0.5
4-3.5=0.5
5-3.5=1.5
6-3.5=2.5

0



As you see the answer is zero. What do I do wrong? and the same with
Y-Y(with a line above). It turns out to be zero. Please help me to tell
how I should do.

Janne





=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Standardizing evaluation scores

2002-01-07 Thread Dennis Roberts

sorry for late reply

ranking is the LEAST useful thing you can do ... so, i would never START 
with simple ranks
any sort of an absolute kind of scale ... imperfect as it is ... would 
generally be better ...

one can always convert more detailed scale values INTO ranks at the end if 
necessary BUT, you cannot go the reverse route

say we have 10 people measured on variable X ... and we end up with no ties 
... so, we get ranks of 1 to 10 ... but, these value give on NO idea 
whatsoever as to the differences amongst the 10

if i had a 3 person senior high school class with cumulative gpas of 4.00, 
3.97, and 2.38 ... the ranks would be 1, 2, and 3 ... but clearly, there is 
a huge difference between either of the top 2 and the bottom ... but, ranks 
give no clue to this at all

so, my message is ... DON'T START WITH RANKS

At 02:11 AM 12/19/01 +, Doug Federman wrote:
I have a dilemma which I haven't found a good solution for.  I work with
students who rotate with different preceptors on a monthly basis.  A
student will have at least 12 evaluations over a year's time.  A
preceptor usually will evaluate several students over the same year.
Unfortunately, the preceptors rarely agree on the grades.  One preceptor
is biased towards the middle of the 1-9 likert scale and another may be
biased towards the upper end.  Rarely, does a given preceptor use the 1-9
range completely.  I suspect that a 6 from an easy grader is equivalent
to a 3 from a tough grader.

I have considered using ranks to give a better evaluation for a given
student, but I have a serious constraint.  At the end of each year, I
must submit to another body their evaluation on the original 1-9 scale,
which is lost when using ranks.

Any suggestions?

--
It has often been remarked that an educated man has probably forgotten
most of the facts he acquired in school and university. Education is what
survives when what has been learned has been forgotten.
- B.F. Skinner New Scientist, 31 May 1964, p. 484




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Looking for some datasets

2002-01-07 Thread Dennis Roberts

some minitab files and other things are here

http://roberts.ed.psu.edu/users/droberts/datasets.htm



_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Excel vs Quattro Pro

2002-01-07 Thread Dennis Roberts

i don't know the answer to this but ... i have a general question with 
regards to using spreadsheets for stat analysis

why? ... why do we not help our students and encourage our students to use 
tools designed for a task ... rather than substituting something that may 
just barely get us by?

we don't ask stat packages to do what spreadsheets were designed to do ... 
why the reverse?

just because packages like excel are popular and readily available ... does 
not therefore mean that we should be recommending it (or them) to people 
for statistical analysis

it's like telling people that notepad will be sufficient to do all your 
word processing needs ...

At 04:56 PM 1/7/02 -0600, Edward Dreyer wrote:
Does anyone know if Quattro Pro suffers the same statistical problems as 
Excel?

Cheers.  ECD
___

Edward C. Dreyer
Political Science
The University of Tulsa






=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: Excel vs Quattro Pro

2002-01-07 Thread dennis roberts

most stat packages have nothing to do with programming anything ... you 
either use simple commands to do things you want done (like in minitab ... 
mtb correlation 'height' 'weight') or, select procedures from menus and 
dialog boxes

At 12:27 AM 1/8/02 +, Kenmlin wrote:
 i don't know the answer to this but ... i have a general question with
 regards to using spreadsheets for stat analysis

Many students are computer illiterate and it might be easier to teach them how
to use the spreadsheet than a formal programming language.




=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
   http://jse.stat.ncsu.edu/
=



=
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at
  http://jse.stat.ncsu.edu/
=



Re: ANOVA = Regression

2001-12-11 Thread Dennis Roberts

of course, the typical software program, when doing regression analysis ... 
prints out a summary ANOVA table ... so, there is one place to start ...


At 10:52 AM 12/11/01 -0500, Wuensch, Karl L wrote:
For a demonstration of the equivalence of regression and traditional ANOVA,
just point your browser to
http://core.ecu.edu/psyc/wuenschk/StatHelp/ANOVA=reg.rtf.

-Original Message-
From:   Stephen Levine [mailto:[EMAIL PROTECTED]]
Sent:   Tuesday, December 11, 2001 3:47 AM
To: Karl L. Wuensch
Subject:Re: When does correlation imply causation?

Hi
You wrote
 Several times I have I had to explain to my colleagues that two-group t
tests and ANOVA are just special cases of correlation/regression analysis.
I can see what you mean - could you please proof it - I read, in a pretty
good text, that the results are not necessarily the same!
Cheers
S.




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When to Use t and When to Use z Revisited

2001-12-10 Thread Dennis Roberts

At 04:14 AM 12/10/01 +, Jim Snow wrote:
Ronny Richardson [EMAIL PROTECTED] wrote in message
[EMAIL PROTECTED]">news:[EMAIL PROTECTED]...

  A few weeks ago, I posted a message about when to use t and when to use z.

I did not see the earlier postings, so forgive me if I repeat advice already
given.:-)

 1. The consequences of using the t distribution instead of the normal
distribution for sample sizes greater than 30 are of no importance in
practice.

what's magical about 30? i say 33 ... no actually, i amend that to 28

 2. There is no good reason for statistical tables for use in practical
analysis of data to give figures for t on numbers of degrees of freedom over
30 except that it makes it simple to routinely use one set of tables when
the variance is estimated from the sample.

with software, there is no need for tables ... period!


 3. There are situations where the error variance is known. They
generally arise when the errors in the data arise from the use of a
measuring instrument with known accuracy or when the figures available are
known to be truncated to a certain number of decimal places. For example:
 Several drivers use cars in a car pool. The distance tavelled on each
trip by a driver is recorded, based on the odometer reading. Each
observation has an error which is uniformly distributed in (0,0.2). The
variance of this error is (0.2)^2)/12  = .00  and standard deviation
0.0578  . To calculate confidence limits for the average distance travelled
by each driver, the z statistic should be used.

this is pure speculation ... i have yet to hear of any convincing case 
where the variance is known but, the mean is not


_____
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When to Use t and When to Use z Revisited

2001-12-10 Thread Dennis Roberts

At 03:42 PM 12/10/01 +, Jerry Dallal wrote:
Dennis Roberts wrote:

  this is pure speculation ... i have yet to hear of any convincing case
  where the variance is known but, the mean is not

A scale (weighing device) with known precision.

as far as i know ... knowing the precision is expressed in terms of ... 
'accurate to within' ... and if there is ANY 'within' attached ... then 
accuracy for SURE is not known





=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-12-07 Thread Dennis Roberts

At 08:08 PM 12/7/01 +, J. Williams wrote:
On 6 Dec 2001 11:34:20 -0800, [EMAIL PROTECTED] (Dennis Roberts) wrote:

 if anything, selectivity has decreased at some of these top schools due to
 the fact that given their extremely high tuition ...


i was just saying that IF anything had happened ... that it might have gone 
down ... i was certainly not saying that it had ...

but i do think that it could probably not get too much more selective ... 
so it probably has sort of stayed where it has over the decades ... so if 
grade inflation has occurred there it would not likely be due to an 
increased smarter incoming class


_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Stat question

2001-12-06 Thread Dennis Roberts

the reality of this is ... sometimes getting notes from other students is 
helpful ... sometimes it is not ... there is no generalization one can make 
about this

most student who NEED notes are not likely to ask people other than their 
friends ... and, in doing so, probably know which of their friends they 
have the best chance of getting good notes from ... (at least READABLE!) 
...even lazy students are not likely to ask for notes from people that even 
THEY know are not going to be able to do them any good

but i don't think we can say anything really systematic about this activity 
other than, sometimes it helps ... sometimes it does not help

At 06:24 PM 12/5/01 -0800, Glen wrote:
Jon Miller [EMAIL PROTECTED] wrote in message 
  You can ask the top students to look at their notes, but you should be 
 prepared
  to find that their notes are highly idiosyncratic.  Maybe even unusable.

Having seen notes of some top students on a variety of occasions
(as a student and as a lecturer), that certainly does happen
sometimes. But just about as likely is to find a set of notes that
are actually better than the lecturer would prepare themselves.

Glen


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-12-06 Thread Dennis Roberts

generally speaking, it is kind of difficult to muster sufficient evidence 
that the amount of grade inflation that is observed ... within and across 
schools or colleges ... is due to an  increase in student ability

i find it difficult to believe that the average ability at a place like 
harvard has gone up ... but if so, very much over the years ...

if anything, selectivity has decreased at some of these top schools due to 
the fact that given their extremely high tuition ... they need to keep 
their dorms full and, making standards higher and higher would have the 
opposite effect on keep dorms filled





At 11:58 AM 12/6/01 -0500, Rich Ulrich wrote:
Just in case someone is interested in the Harvard instance
that I mentioned -- while you might get the article from a newsstand
or a friend --

On Sun, 02 Dec 2001 19:19:38 -0500, Rich Ulrich [EMAIL PROTECTED]
wrote:

=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Experimental Correlation Coefficients

2001-12-06 Thread Dennis Roberts

i would say that karl has demonstrated that IF we know conditions of 
manipulation or not ... we can have a better or lesser idea of what (if 
anything) impacted (caused?)  what

that i will grant him

to argue that r or eta has anything to do with this ... i would 
respectfully disagree

they are just byproducts of our manipulations ...

one cannot equate the byproducts with THE manipulations ..

At 03:23 PM 12/6/01 -0500, Wuensch, Karl L wrote:
 My experimental units are 100 classrooms on campus.  As I walk into
each room I flip a perfectly fair coin in a perfectly fair way to determine
whether I turn the room lights on (X = 1) or off (X = 0).  I then determine
whether or not I can read the fine print on my bottle of smart pills (Y = 0
for no, Y = 1 for yes).  From the resulting pairs of scores (one for each
classroom), I compute the phi coefficient (which is a Pearson r computed
with dichotomous data).  Phi = .5.  I test and reject the null hypothesis
that phi is zero in the population (using chi-square as the test statistic).
Does correlation (phi is not equal to zero) imply causation in this case?
That is, can I conclude that turning the lights on affects my ability to
read fine print?

could be here that if you did not have glasses ... you could not have read 
anything with or without light ... and, since you did have glasses ... the 
r you get is because of the implicit interaction between light or not, and 
glasses or not



  ~~~
Karl L. Wuensch, Department of Psychology,
East Carolina University, Greenville NC  27858-4353
Voice:  252-328-4102 Fax:  252-328-6283
mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
http://core.ecu.edu/psyc/wuenschk/klw.htm
http://core.ecu.edu/psyc/wuenschk/klw.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When does correlation imply causation?

2001-12-05 Thread Dennis Roberts

correlation NEVER implies causation ...
and i agree with mike totally

At 09:01 AM 12/5/01 -0600, Mike Granaas wrote:


We really need to emphasize over and over that it is the manner in which
you collect the data and not the statistical technique that allows one to
make causal inferences.

Michael

 
 
 
 
  
 
 
  Karl L. Wuensch, Department of Psychology,
  East Carolina University, Greenville NC  27858-4353
  Voice:  252-328-4102 Fax:  252-328-6283
  [EMAIL PROTECTED]
  http://core.ecu.edu/psyc/wuenschk/klw.htm
 
 
 

***
Michael M. Granaas
Associate Professor[EMAIL PROTECTED]
Department of Psychology
University of South Dakota Phone: (605) 677-5295
Vermillion, SD  57069  FAX:   (605) 677-6604
***
All views expressed are those of the author and do not necessarily
reflect those of the University of South Dakota, or the South
Dakota Board of Regents.



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When does correlation imply causation?

2001-12-05 Thread Dennis Roberts

At 07:36 AM 12/5/01 -0500, Karl L. Wuensch wrote:


   Accordingly, I argue that correlation is a necessary but not a 
 sufficient condition to make causal inferences with reasonable 
 confidence.  Also necessary is an appropriate method of data 
 collection.  To make such causal inferences one must gather the data by 
 experimental means, controlling extraneous variables which might confound 
 the results.  Having gathered the data in this fashion, if one can 
 establish that the experimentally manipulated variable is correlated with 
 the dependent variable (and that correlation does not need to be linear), 
 then one should be (somewhat) comfortable in making a causal 
 inference.  That is, when the data have been gathered by experimental 
 means and confounds have been eliminated, correlation does imply causation.

the problem with this is ... does higher correlation mean MORE cause? lower 
r mean LESS cause?

in what sense can think of cause being more or less? you HAVE to think that 
way IF you want to use the r value AS an INDEX MEASURE of cause ...

personally, i think it is dangerous in ANY case to say that r = cause ...

if you can establish that as A goes up ... so does B ... where you 
manipulated A and measured B ... (or vice versa) ... then it is fair to say 
that the causal connection THAT IS IMPLIED BECAUSE OF THE WAY THE DATA WERE 
MANIPULATED/COLLECTED also has a concomitant r ... BUT, i think one still 
needs to be cautious when then claiming that the r value itself is an 
indicant OF cause










   So why is it that many persons believe that one can make causal 
 inferences with confidence from the results of two-group t tests and 
 ANOVA but not with the results of correlation/regression techniques.  I 
 believe that this delusion stems from the fact that experimental research 
 typically involves a small number of experimental treatments that data 
 from such research are conveniently evaluated with two-group t tests and 
 ANOVA.  Accordingly, t tests and ANOVA are covered when students are 
 learning about experimental research.  Students then confuse the 
 statistical technique with the experimental method.  I also feel that the 
 use of the term correlational design contributes to the problem.  When 
 students are taught to use the term correlational design to describe 
 nonexperimental methods of collecting data, and cautioned regarding the 
 problems associated with inferring causality from such data, the students 
 mistake correlational statistical techniques with correlational data 
 collection methods.  I refuse to use the word correlational when 
 describing a design.  I much prefer nonexperimental or observational.



   In closing, let me be a bit picky about the meaning of the word 
 imply.  Today this word is used most often to mean to hint or to 
 suggest rather than to have as a necessary part.  Accordingly, I argue 
 that correlation does imply (hint at) causation, even when the 
 correlation is observed in data not collected by experimental means.  Of 
 course, with nonexperimental models, the potential causal explanations of 
 the observed correlation between X and Y must include models that involve 
 additional variables and which differ with respect to which events are 
 causes and which effects.

--
Karl L. Wuensch, Department of Psychology,
East Carolina University, Greenville NC  27858-4353
Voice:  252-328-4102 Fax:  252-328-6283
mailto:[EMAIL PROTECTED][EMAIL PROTECTED]
http://core.ecu.edu/psyc/wuenschk/klw.htm

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: When does correlation imply causation?

2001-12-05 Thread Dennis Roberts

perhaps the problem here is with the word ... cause

say i put in a column some temps in F ... then use the F to C formula ... 
and get the corresponding C values ...

then i do an r between the two and find 1.0

now, is the formula the cause of the r of 1?

maybe we might see it as a cause but ... then again ... if i had Cs and 
changed to Fs ... i get the same thing ...

now, what if we have 3 dosage levels (we manipulate) of 10, 20, and 30 ... 
and we find that on the criterion ... we get a graph of


Plot


  -  *
  1.50+
  -
  res1-
  - *
  -
  1.00+
  -
  -*
  -
+-+-+-+-+dosage
 12.0  16.0  20.0  24.0  28.0

and an r = 1

MTB  corr c1 c2

Correlations: dosage, res1


Pearson correlation of dosage and res1 = 1.000
P-Value = *

but, for another set of data we get

MTB  plot c3 c1

Plot


  res2-  *
  -
  -
  1.00+
  -
  - *
  -
  -
  0.80+*
  -
+-+-+-+-+dosage
 12.0  16.0  20.0  24.0  28.0

MTB  corr c1 c3

Correlations: dosage, res2


Pearson correlation of dosage and res2 = 0.982
P-Value = 0.121

and an r = .982

the fact is that we could find and equation (not linear) where the dots 
about would be hit perfectly ... ie, find another model where there is no 
error

but, the r is not 1 ... so, does that mean that some other factor is 
detracting from our cause of Y?

the fact that the pattern is NOT linear (to which r is a function) ... is 
not a detraction ...

if the model fails to find an r = 1, then that does not mean that there is 
lack of perfect cause, whatever that means ... only that the model does 
not detect it

so, in that sense, lower rs do not necessarily mean that there are other 
errors or extraneous factors that enter ... i would not call using the 
wrong model an extraneous factor




At 02:12 PM 12/5/01 -0500, Wuensch, Karl L wrote:
Dennis warns the problem with this is ... does higher correlation mean MORE
cause? lower r mean LESS cause?
in what sense can think of cause being more or less? you HAVE to think that
way IF you want to use the r value AS an INDEX MEASURE of cause ...

Dennis is not going to like this, since he has already expressed a disdain
of r-square, omega-square, and eta-square like measures of the strength of
effect of one variable on another, but here is my  brief reply:

R-square tells us to what extent we have been able to eliminate, in our data
collection procedures, the contribution of other factors which influence the
dependent variable.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: When does correlation imply causation?

2001-12-05 Thread Dennis Roberts

what if we see, from the side, one person on the right swing his/her fist 
... and, as the fist arrives at what appears to be the impact point of the 
face of a person on the left ... that IMMEDIATELY the person on the left 
falls backwards and down

now, we do this over and over again ... and observe the same result

we want to say that the impact to the FACE ... CAUSED the person on the 
left to fall down

but, did it? in a sense, this is like a perfect correlation in that ... 
when the person swung to the RIGHT .. the person on the left NEVER fell 
backwards and down

with an r = 1, karl would say that this has removed ALL extraneous factors 
from the possibility of an explanation for why the person fell ... or did 
not fall

however, what is UNBEKNOWNST to the viewer ... there was a clear panel 
between the two people ... thrower of the punch and, the one on the left 
... and, each time the person threw a punch ... and it LOOKED like the 
punch TO the person caused a fall down, ... in actuality there was an 
electronic triggering mechanism that ... when the panel was activated ... 
lead to some electrical shock being applied to the person on the left ... 
that made the person fall down

it could be that if you viewed this from the angle of both people ... 90 
degrees rotated ... you could see that there was a clear separation of 2 or 
more feet between where the punch hit the clear panel and ... where the 
face was leaning up against the panel

so, the punch did NOT directly touch the face ... and in that sense, could 
not have caused the person to fall backwards and down

even though these may be the facts ... the sideways view clearly has 
missed that there were OTHER things in the chain of events ... that LEAD to 
the person falling backwards and down

r values ... even perfect ones ... are in an impossible statistical 
position to say anything about cause ... or, more importantly, what causes what

the r that goes along with some manipulative procedure that varies X and 
produces some corresponding Y change ... is just a tag along statistic ... 
and may not really (even when perfect) suggest anything about HOW X, when 
manipulated, PRODUCED the Y change that we see ...

thus, the use of r in this case as an index of how MUCH X CAUSES Y ... 
is  a statistical stretch






_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



glass and stanley

2001-12-02 Thread Dennis Roberts

does anyone know where i might get a copy of the 1970 book, stat. methods 
in ed. and psy. ... by gene glass and julian stanley?

mine seems to have disappeared and, i would like to retrieve a copy

thanks for any leads

PLEASE SEND ME A PERSONAL NOTE IF YOU HAVE ANY INFO ... and not to the list

emailto:[EMAIL PROTECTED]



_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Stat question

2001-12-01 Thread dennis roberts

At 06:13 PM 12/1/01 -0500, Stan Brown wrote:
Jon Miller [EMAIL PROTECTED] wrote in sci.stat.edu:
 
 Stan Brown wrote:
 
  I would respectfully suggest that the OP _first_ carefully study the
  textbook sections that correspond to the missed lectures, get notes from
  a classmate
 
 This part is of doubtful usefulness.

Doubtful? It is of doubtful usefulness to get notes from a
classmate and study the covered section of the textbook? Huh?

perhaps doubtful IF the students OP asked to look at were terrible students 
who took terrible notes ... and/or ... OP when reading the text could not 
make anything of it ...

but, those are two big ifs

usually, students won't ask to see the notes of students whom they know are 
not too swift ... and, also ... usually students who read the book do get 
something out of it ... maybe not enough

the issue here is ... it appeared (though we have no proof of this) that 
the original poster did little, if anything, on his/her own ... prior to 
posting a HELP to the list

stan seemed to be reacting to that assumption and, i don't blame him


--
Stan Brown, Oak Road Systems, Cortland County, New York, USA
   http://oakroadsystems.com/
My theory was a perfectly good one. The facts were misleading.
-- /The Lady Vanishes/ (1938)


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: experimental design(doubt)?

2001-12-01 Thread dennis roberts

i think this points out that it is hard to really give good responses 
sometimes when all the details are not known ... in this case, we really 
don't have sufficient information on HOW samples were selected and 
assigned, METHODS and orders that items were heated and then porcelain 
applied, and on and on

for it to be totally randomized, we would have to have something akin to 
... having 72 samples ... assigning them to temp and porcelain type first 
... then executing this design in that order ...

could be that sample 1 got 430 degrees and type b, sample 2 might have 
gotten temp 700 with porcelain type b, and so on

but, we probably know that is NOT what happened ... because that would have 
created implementation problems

we know in factories ... there are runs of different items at different 
times ... they might have a run of X from 8AM to NOON, then there is a 
transition period before Y gets done from 1PM to 5PM ...

in this instance, it probably was the case that all 24 were heated to the 
first temp ... then when that was all done, the oven was revved up to the 
next higher temp and then the next 24 were heated ... and a final heat up 
to the last temp saw the final 24 done

so, this is not exactly a totally randomized plan ... since there could 
have been some systematic difference from one batch to the other

we also don't know how the porcelain was applied ... it might have been 
that after all were heated to temp 1 ... then the first 12 that came out of 
the oven were given porcelain type A ... since this was easier to do ... 
then the last 12 got (after the change over) porcelain type B

if either temp or type of porcelain made a BIG difference, these procedural 
details probably don't make a hill of beans of difference but, of course, 
if the impacts (though maybe real) were very small, then some systematic 
error might make a difference

as i said ... we just don't know

however, i think trying to give the design the proper NAME is really not 
that important ... the real important matter is whether the implementation 
of the plan was sufficiently close enough to a fully randomized design that 
... an analysis according to that design would be satisfactory in this case

bottom line is: snippets of information given to the list ... does not 
necessarily allow us to field the ensuing inquiries ... and, it is better 
to probe more about methods and procedures first ... then to rush off with 
some analysis conclusion

BUT WE DO IT ANYWAY!





If the samples have been treated independently, that is each sample is
individually raised to the randomly assigned temperature and
subsequently treated with the assigned porcelain type, the design is a
completely randomized design. Any application effects (including
possible deviations of supposed temperatures and irregularities during
the whole proces of heating and subsequent cooling) will contribute to
the random error of the observations. But when all samples of the same
temperature treatment are simultaneously put in the furnace and
treated as one batch the situation is different. In that case
application effects (whose existence and magnitude is not known
generally) are confounded with the effects of the temperature
treatment. In my comment I supposed and stated that probably this was
the situation at hand. I have to admit that the original message is
not entirely clear on this point.

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: experimental design(doubt)?

2001-11-30 Thread Dennis Roberts

At 03:11 PM 11/30/01 -0200, Ivan Balducci wrote:
Hi, members:
Please, I am hoping to get some clarification on this problem:
In my University, (I work in Dental School, in Brazil),
a dentist brought to me your data:
Experimental Unit:
cilindrical shape of Titanium pure ( 72 samples): diameter: 4mmm; height: 5mm
submeted to Shear Test (Instrom)
..
The Ti were to heating in furnace:
24 samples to 430ºC;
24 samples to700ºC
24 samples to and 900ºC
..
post hoc
...
12 samples (from 430ºC) received porcelain type A
12 samples (from 430ºC) received porcelain type B
...
12 samples (from 700ºC) received porcelain type A
12 samples (from 700ºC) received porcelain type B
.
12 samples (from 900ºC) received porcelain type A
12 samples (from 900ºC) received porcelain type B
... 

Objectives:
Effect Interaction between the variables: Temperature and Porcelain
on Shear Data;
TIA,
ivan
My question is: She made an Split-Plot design ?
Whole plot: Temperature
Split: Porcelain.

looks like a simple randomized design to me ... in effect, you have 
selected 12 at random and raised to 430 AND gave porcelain type A ...

and the other 5 combinations...

3 by 2 design ... fully randomized ...

unless i am missing something




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Interpreting p-value = .99

2001-11-29 Thread Dennis Roberts

forget the statement of the null

build a CI ... perhaps 99% (which would correspond to your .01 sig. test) ...

let that help to determine if the claim seems reasonable or not

in this case ... p hat = .85 .. thus q hat = .15

stan error of a proportion (given SRS was done) is about

stan error of p hats = sqrt ((p hat * q hat) / n) = sqrt (.85 * .15 / 200) 
= about .025

approximate 99% CI would be about p hat +/-  2.58 * .025 = .85 +/- .06

CI would be about .79 to .91 ... so, IF you insist on a hypothesis test ... 
retain the null

personally, i would rather say that the pop. proportion might be between 
(about) .79 to .91 ...

doesn't hold me to .9

problem here is that if you have opted for .05 ... you would have rejected 
... (just barely)

At 02:39 PM 11/29/01 -0500, you wrote:
On a quiz, I set the following problem to my statistics class:

The manufacturer of a patent medicine claims that it is 90%
effective(*) in relieving an allergy for a period of 8 hours. In a
sample of 200 people who had the allergy, the medicine provided
relief for 170 people. Determine whether the manufacturer's claim
was legitimate, to the 0.01 significance level.

(The problem was adapted from Spiegel and Stevens, /Schaum's
Outline: Statistics/, problem 10.6.)


I believe a one-tailed test, not a two-tailed test, is appropriate.
It would be silly to test for effectiveness differs from 90% since
no one would object if the medicine helps more than 90% of
patients.)

Framing the alternative hypothesis as the manufacturer's claim is
not legitimate gives
 Ho: p = .9; Ha: p  .9; p-value = .0092
on a one-tailed t-test. Therefore we reject Ho and conclude that the
drug is less than 90% effective.

But -- and in retrospect I should have seen it coming -- some
students framed the hypotheses so that the alternative hypothesis
was the drug is effective as claimed. They had
 Ho: p = .9; Ha: p  .9; p-value = .9908.

Now as I understand things it is not formally legitimate to accept
the null hypothesis: we can only either reject it (and accept Ha) or
fail to reject it (and draw no conclusion). What I would tell my
class is this: the best we can say is that p = .9908 is a very
strong statement that rejecting the null hypothesis would be a Type
I error. But I'm not completely easy in my mind about that, when
simply reversing the hypotheses gives p = .0092 and lets us conclude
that the drug is not 90% effective.

There seems to be a paradox: The very same data lead either to the
conclusion the drug is not effective as claimed or to no
conclusion. I could certainly tell my class: if it makes sense in
the particular situation, reverse the hypotheses and recompute the
p-value. Am I being over-formal here, or am I being horribly stupid
and missing some reason why it _would_ be legitimate to draw a
conclusion from p=.9908?

--
Stan Brown, Oak Road Systems, Cortland County, New York, USA
   http://oakroadsystems.com
My reply address is correct as is. The courtesy of providing a correct
reply address is more important to me than time spent deleting spam.


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Q and t

2001-11-28 Thread Dennis Roberts

in what way are the Q values (critical) and t ... connected?

someone pointed out that Q = t * sqrt 2 ... for example, doing a tukey 
follow up test ... with 3 means and 27 df for mserror ... the critical (.95 
one tail) value for Q is about 3.51 (note: the mtb output shows this cv of 
3.51)

now, that means that t * sqrt 2 = 3.51 ... or, t will equal about 2.48 in 
this case

in minitab, doing a onew anova ... or glm ... with tukey as the paired 
comparison method ... the 95% CI for the paired difference in means (for 
the problem i was working with) did go from (mean difference) +/- 2.48 * 
stan error of the difference

my question is ... how would we go to the standard t table ... and find a t 
value like 2.48 in a situation with the df value we have above ... 27 ... 
or knowing that we have 3 means to compare?

that is ... how would someone know how to construct this 95% CI ... that 
mtb prints out when doing tukey follow up tests? ...



_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students: A Statistical Perspective

2001-11-28 Thread Dennis Roberts

At 01:35 PM 11/28/01 -0600, jim clark wrote:
Hi

On Tue, 27 Nov 2001, Thom Baguley wrote:
  I'd argue that they probably aren't that independent. If I ask three
  questions all involving simple algebra and a student doesn't
  understand simple algebra they'll probably get all three wrong. In
  my experience most statistics exams are better represented by a
  bimodal (possibly a mix of two skewed normals) than a normal
  distribution. Essay based exams tend to end up with a more unimodal
  distribution (though usually still skewed).

The distribution of grades will depend on the distribution of
difficulties of the items, one of the elements examined by
psychometrists in the development of professional-quality
assessments.

well, not exactly ... it depends on a joint function of how hard items turn 
OUT to be AND, where i set the cut scores for grades

items can be real difficult ... but still exhibit some spread .. hence my 
distribution of grades may or may not exhibit some spread depending on 
where i set the A, B, etc. points

item difficulties will determine (usually) the general SHAPE of the 
distribution of SCORES ... but grades are on top of scores and do NOT have 
to conform to the shape of the distribution of scores

unless your semantics was  equating the term grades with the term scores ...




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Meas. Position

2001-11-26 Thread Dennis Roberts

The Educational Measurement position (Assistant/Associate Professor) at 
Penn State University has been officially posted at the College of 
Education website ...

If you have any interest in this or know of someone who might be, please 
pass along the link below

Thanks

http://www.ed.psu.edu/employment/edpsymeas.asp

This should also be posted at AERA.NET shortly

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: anyone using datadesk ?

2001-11-23 Thread dennis roberts

datadesk is now linked on minitabs homepage ...as is minitab being linked 
on datadesks homepage ...

At 11:36 AM 11/23/01 +, Frank Mattes wrote:
Hi
I discovered datadesk (www.datadesk.com), a application which allows
graphical exploration of your data.
I'm wondering if anyone is using this software

Thank

Frank Mattes


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: CIs for adjusted rates

2001-11-21 Thread Dennis Roberts

At 08:35 AM 11/21/01 -0600, Scheltema, Karen wrote:
I have complication rates for a given procedure.  I was thinking of using
indirect standardization as a method of risk adjustment given that some
doctors see more complex patients.  What I can't figure out is how I would
go about calculating a 95% CI after the risk adjustment.  Any pointers would
be greatly appreciated.

why not define a group of doctors or cases as ... less complex patients .. 
and more complex patients ... and do the confidence intervals separately 
(you might do one overall interval to see how it compares to the two 
separate ones ... perhaps there is little difference)

one problem i do see is that ... some doctors deal with both so, the two 
classes above (less complex, more complex) are not totally independent


Karen Scheltema, M.A., M.S.
Statistician
HealthEast
Research and Education
1700 University Ave W
St. Paul, MN 55104
(651) 232-5212 (phone)
(651) 641-0683 (fax)
[EMAIL PROTECTED]




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



best inference

2001-11-21 Thread Dennis Roberts

on this near holiday ... at least in the usa ... i wonder if you might 
consider for a moment:

what is the SINGLE most valuable concept/procedure/skill (just one!) ... 
that you would think is most important  when it comes to passing along to 
students studying inferential statistics

what i am mainly looking for would be answers like:

the notion of 

being able to do __

that sort of thing

something that if ANY instructor in stat, say at the introductory level 
failed to discuss and emphasize ... he/she is really missing the boat and 
doing a disservice to students



_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: When Can We Really Use CLT Student t

2001-11-21 Thread Dennis Roberts

At 12:49 PM 11/21/01 -0500, Ronny Richardson wrote:
As I understand it, the Central Limit Theorem (CLT) guarantees that the
distribution of sample means is normally distributed regardless of the
distribution of the underlying data as long as the sample size is large
enough and the population standard deviation is known.

nope ... clt says nothing of the kind
it says that regardless of the shape of the target population ... as n 
increases, the shape of the sampling distribution of means is better and 
better APPROXIMATED by the normal distribution

that is, even if the target population is quite different from normal ... 
if we take decent sized samples ... we can say and not be TOO wrong that 
the sampling distribution of means looks something like a normal ...

here is a quick simulation taking samples of n=50 (based on 1 samples) 
from a chi square distribution with 1 df

.
  ..::..
:.
  ..
 .::..
   .::.
 ..::..
   .: .
  +-+-+-+-+-+---C51
   0.30  0.60  0.90  1.20  1.50  1.80

even though the chi square distribution is radically + skewed, the sampling 
distribution looks pretty darn close to a normal distribution ... but it 
never will be exactly one ...


it does NOT say that it will GET to and BECOME a normal distribution

if the population is not normal ... the sampling distribution will not be 
normal regardless of n ... but, it could be that your EYES could not tell 
the difference


It seems to me that most statistics books I see over optimistically invoke
the CLT not when n is over 30 and the population standard deviation is
known but anytime n is over 30. This seems inappropriate to me or am I
overlooking something?

you are mixing two metaphors ...

if we know the sd of the population ... then we know the real sampling 
error ... ie, standard error of the mean ... if we do NOT know the 
population sd, and substitute our estimate of that from the sample, then we 
are only estimating the standard error of the mean

thus ... knowing or not knowing the population sd helps us to know or only 
to estimate the real standard error ... but this is unconnected with shape 
of sampling distribution

shape of sampling distribution is partly a function of shape of population 
AND random sample size ...


When the population standard deviation is not know (which is almost all the
time) it seems to me that the Student t (t) distribution is more
appropriate. However, t requires that the underlying data be normal, or at
least not too non-normal. My expectations is that most data sets are not
nearly normal enough to make using t appropriate.

So, if we do not know the population standard deviation and we cannot
assume a normal population, what should we be doing-as opposed to just
using the CLT as most business statistics books do?

Ronny Richardson


Ronny Richardson


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



part. SS

2001-11-19 Thread Dennis Roberts

i have put up a little worked out example of the partitioning of SS in a 
simple 3 group situation ... for ANOVA ... with a diagram of where the 
components come from ...

http://roberts.ed.psu.edu/users/droberts/introstat/sspart1.png

this might be helpful to some ...

if you want to print ... you should do it in the landscape (horizontal) mode

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: biostatistics careers

2001-11-19 Thread Dennis Roberts

At 03:08 PM 11/19/01 +, A.J. Rossini wrote:
  BW == Bruce Weaver [EMAIL PROTECTED] writes:

 BW On Sun, 18 Nov 2001, Stan Brown wrote:

  What _is_ biostatistics, anyway? A student asked me, and I
  realized I have only a vague idea.


 BW There was a thread on biostatistics versus statistics a
 BW couple years ago, I think, but I was unable to find it at
 BW google groups.  Maybe someone out there saved some of it.


But it's much easier than that.  Biostatistics is simply statistics
(design, descriptive, and inferential) applied to medical, basic
biology, and public health problems.

well, one difference in bio stat is a strong emphasis on probability sorts 
of problems ...


_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Maximized lambda4

2001-11-19 Thread Dennis Roberts

At 01:49 PM 11/19/01 -0500, Wuensch, Karl L wrote:
 Callender and Osburn (Educational and Psychological Measurement,
1977, 37, 819-825) developed a method for estimating maximized lambda4, the
greatest split-half reliability coefficient among all possible split halves
for a scale.  The method is quite tedious to do by hand, and the authors
provided a FORTRAN program for accomplishing it.  Not having a FORTRAN
compiler, I'm considering writing an SAS program (IML) to do this, but don't
want to waste my time reinventing the wheel if someone else has already
written an SAS or SPSS program for this purpose.  If you are aware of any
such program, please advise me.  Thanks.

 By the way, Callender and Osburn's work indicates that maximized
lambda4 is a much better estimate of a test's true reliability than is
lambda3 (coefficient alpha).

how do we know what a test's true reliability is?



  ~~~
Karl L. Wuensch, Department of Psychology,
East Carolina University, Greenville NC  27858-4353
Voice:  252-328-4102 Fax:  252-328-6283
mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
http://core.ecu.edu/psyc/wuenschk/klw.htm
http://core.ecu.edu/psyc/wuenschk/klw.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students

2001-11-18 Thread dennis roberts
 action that 
is taken, etc. ... we need to stress both process and accuracy ...

i sure hope that the bombardier is not just graded on process




  ___
  dennis roberts, educational psychology, penn state university
  208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
  http://roberts.ed.psu.edu/users/droberts/drober~1.htm
 
  =
  Instructions for joining and leaving this list and remarks about
  the problem of INAPPROPRIATE MESSAGES are available at
http://jse.stat.ncsu.edu/
  =

--
Roy St. Laurent
Mathematics  Statistics
Northern Arizona University
http//odin.math.nau.edu/~rts

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Introductory Statistics text

2001-11-18 Thread dennis roberts

At 07:34 PM 11/18/01 -0800, Melady Preece wrote:
I am looking for a new and improved Statistics text for an introductory (3rd
year) stats course for psychology majors...I would welcome any
suggestions/reviews, etc.

Melady Preece

improved over what? what are you using? what don't you like about it? is 
software used and if so, does the book you are using (or would like to use) 
help out in this?



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students

2001-11-18 Thread dennis roberts



the general problems evaluating students are how much time do you have for 
(say) exams, what can be reasonably  expected that students will be able to 
do with that amount of time, what content can you examine on, and ... what 
sort of formats do you opt for with your exams

in statistics, let's say you have covered several inferential topics from 
midterm to the end ... maybe some difference in means situations using t 
tests and building CIs, perhaps something about tests and CIs about 
proportions, and then a unit on ANOVA

(note: the above are just examples)

how much can you have them do ... in an open ended way ... in an hour?

for example, i can easily visualize having very small sets of data ... that 
they work with for each of the above ... with appropriate distribution 
tables at hand ... giving written explanations of their results ... but, is 
it realistic to expect them to do something from all of these in one hour? 
i really doubt it

so, in the final analysis, ANY test (no matter what the format) is just a 
sample of all the things you could ask them to do ... thus, no matter what 
you do and what they show ... you are still left with many unanswered 
questions about their knowledge and skill

i think it would be fair to say that in having students work problems and 
show their work ... you do get a better idea of their knowledge of THAT 
but, generally, you are not able to have as widespread coverage of the 
material you have covered since the last test ... than perhaps with some 
recognition item approach ... where you can cover more but, clearly, you 
get less information about any specific piece of knowledge or skill

ultimately, it is a tradeoff ... and, as i think someone else mentioned, if 
the instructor is also strapped with large classes (which is so common 
these days) ... practical considerations enter that weigh perhaps more 
heavily than pedagogic best practice


dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students

2001-11-16 Thread Dennis Roberts

At 08:56 AM 11/16/01 -0700, Roy St Laurent wrote:
It's not clear to me whether recent posters are serious about these
examples, but
I will reiterate my previous post:

For most mathematics / statistics examinations, the answer to a
question is the
*process* by which the student obtains the incidental final number or
result.
The result itself is most often just not that important to evaluating
students'
understanding or knowledge of the subject.  And therefore an unsupported

or lucky answer is worth nothing.

the problems with the above are twofold:

1. this assumes that correct answers are NOT important ... (which believe 
me if you are getting change from a cashier, etc. etc. ... ARE ... we just 
cannot say that knowing the process but not being able to come up with a 
correct answer ... = good performance)

2. that answers without any OTHER supplied information on the part of the 
examinee can't be taken as knowledge when, it (sometimes) can be

what if you asked on an exam ... the following:

1. what is the mean of 10, 9, 8, 8 and 7? _

2. what is the mean of 27, 23, 19, 17 and 16? 

3. what is the mean of 332, 234, 198, 239, and 200? _

4. what is the mean of 23.4, 19.8, 23.1, 19.0, and 26.4? _

and, for each of 1 to 4 ... they put down in the blanks, the correct answers

would you be willing to say that they know how to calculate the mean ... 
ie, they know the process that is needed (and can implement it)?

i think you would EVEN though there is no other supporting process 
information given by the examinee

so, the statement that no credit should be given when there is no 
supporting other evidence (ie, the process is shown) ... can't be 
considered necessarily valid

the problem here is NOT that no supporting evidence is given, the problem 
is that with ONLY ONE instance of some given concept/skill that we are 
attempting to assess on the part of the examinee, you are not nearly as 
sure given only ONE CORRECT RESPONSE to one item ... whether it could have 
been an answer because of real knowledge or, just happens to be the right 
answer that was arrived at (luckily for the examinee) through some faulty 
process






_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



diff in proportions

2001-11-15 Thread dennis roberts

in the moore and mccabe book (IPS), in the section on testing for 
differences in population proportions, when it comes to doing a 'z' test 
for significance, they argue for (and say this is commonly done) that the 
standard error for the difference in proportions formula should be a POOLED 
one ... since if one is testing the null of equal proportions, then that 
means your null is assuming that the p*q combinations are the SAME for both 
populations thus, this is a case of pooling sample variances to estimate a 
single common population variance

but since this is just a null ... and we have no way of knowing if the null 
is true (not that we can in any case) ... i don't see any logical 
progression that would then lead one to also assume that the p*q 
combinations are the same in the two populations ... hence, i don't see why 
the pooled variance version of the standard error of a difference in 
proportions formula would be the recommended way to go

in their discussion of differences in means ... they present FIRST the NON 
pooled version of the standard error and that is there preferred way to 
build CIs and do t tests ... though they also bring in later the pooled 
version as a later topic (and of course if we KNEW that populations had the 
same variances, then the pooled version would be useful)

it seems to me that this same logic should hold in the case of differences 
in proportions

comments?

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: diff in proportions

2001-11-15 Thread Dennis Roberts

At 08:51 AM 11/15/01 -0600, jim clark wrote:

The Ho in the case of means is NOT about the variances, so the
analogy breaks down.  That is, we are not hypothesizing
Ho: sig1^2 = sig2^2, but rather Ho: mu1 = mu2.  So there is no
direct link between Ho and the SE, unlike the proportions
example.

would it be correct then to say ... that the test of differences in 
proportions is REALLY a test about the differences between two population 
variances?


Best wishes
Jim


James M. Clark  (204) 786-9757
Department of Psychology(204) 774-4134 Fax
University of Winnipeg  4L05D
Winnipeg, Manitoba  R3B 2E9 [EMAIL PROTECTED]
CANADA  http://www.uwinnipeg.ca/~clark




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: diff in proportions

2001-11-15 Thread dennis roberts

At 08:03 PM 11/15/01 +, Radford Neal wrote:
Radford Neal:

  The difference is that when dealing with real data, it is possible for
  two populations to have the same mean (as assumed by the null), but
  different variances.  In contrast, when dealing with binary data, if
  the means are the same in the two populations, the variances must
  necessarily be the same as well.  So one can argue on this basis that
  the distribution of the p-values if the null is true will be close to
  correct when using the pooled estimate (apart from the use of a normal
  approximation, etc.)

Jerry Dallal:

 But, if the null hypothesis is that the means are the same, why
 isn't(aren't) the sample variance(s) calculated about a pooled
 estimate of the common mean?


An interesting question.


i think what  this shows (ie, these small highly technical distinctions) is 
that ... that most null hypotheses that we use for our array of 
significance tests ... have rather little meaning

null hypothesis testing is a highly overrated activity in statistical work

in the case of differences between two proportions ... the useful question 
is: i wonder how much difference (since i know there is bound to be some 
[even though it could be trivial]) there is between the proportions of A 
population versus B population?

to seek an answer to the real question ... no notion of null has to even be 
entertained


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students

2001-11-15 Thread dennis roberts

would we give full credit for 87/18 = 7/1 ... 8's cancel?



Full marks. As Napoleon used to ask, Is he lucky?.  :)  He/she deserves it.!

  --
John Kane
The Rideau Lakes, Ontario Canada




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Evaluating students

2001-11-14 Thread Dennis Roberts

the problem with any exam ... given in any format ... is the extent to 
which you can INFER what the examinee knows or does not know from their 
responses

in the case of recognition tests ... where precreated answers are given and 
you make a choice ... it is very difficult to infer anything BUT what the 
choice was ... if that choice was keyed as correct and the examinee 
selected it ... first, you have to give credit for it and second ... about 
all you can do is to assume the person understood the item (which may or 
may not be correct of course)

in open ended problems ... where the person has to SUPPLY responses ... the 
instructor is in somewhat of a better position to assess knowledge ... 
assuming that the directions to the examinee were clear as to what he/she 
was to put down ...

however, even here, there are limits

what if i ask the ? ... what is the mean of 34, 56, 29, and 32? and i leave 
space ... and your directions say to round to 1 place and he/she puts 37.8 
in the space ... sort of looks like he/she knew what to do (but you can't 
be positive since, he/she could have added up slightly incorrectly but 
still when rounding, it = 37.8)

but, what if they put 37.2  or 38.1 ... what do you make of it?

now, if WORK were required to be displayed, you might be able to see what 
happened and then grade accordingly ... if the process was totally messed 
up ... no credit but, if you see they added up the 4 values incorrectly 
but, had the right process ... almost full credit but, not total ... they 
DID make some mistake

what if they added correctly but for some unknown reason ... had in the 
back of their mind n-1 and divided by 3? is that = to a mistake of dividing 
by 4 but having added up the 4 numbers wrong?

now, in a mc format ... we might have 37.8 as one of the answers but, other 
options that incorporate some adding up the numbers INcorrectly but 
dividing by 4, adding up the 4 numbers correctly but dividing by n-1, etc.

so, if they don't select the keyed answer ... BUT, they select some other 
option ... then you MIGHT be able to infer some partial knowledge just as 
if you see an incorrect answer and see that they divided by 3 rather than 4

BUT YOU HAVE TO BE CAREFUL DOING THIS ... since you still see no worked out 
evidence of where they might have gone astray

generally, open ended ? that require physical responses to be made off a 
better opportunity to gauge what the examinee knows but ... again ... 
there are real limits

the bottom line here is that no matter how you examine the student, unless 
you do lots of followup probing ... you will be in a difficult position to 
know very precisely what the person knows or does not know

an exam is a very limited sample of behavior from which you are not in a 
very good position to extend your inference about what the examinee was 
really thinking when faced with that item or test

to attempt to infer more is playing a sheer folly game




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re:

2001-11-14 Thread Dennis Roberts

At 07:42 AM 11/14/01 -0800, Carl Huberty wrote:
I, too, prefer closed-book tests in statistical methods courses.  I also 
like short-answer items, some of which may be multiple-choice 
items.  [Please don't gripe that all multiple-choice items assess only 
memory recall; such items, if constructed well, may be very helpful in 
assessing learning!]  I think that a very important aspect of evaluation 
of student performance and knowledge pertains to variability; variability 
in the sense of class performance.  If assessment of student learning does 
not reflect some variability in student performance, there is a very 
serious problem with the assessment process used!

good point ... and, EVEN if every student came in with the same identical 
ability (say math skill) ... there is no way that whatever happens in the 
course will equally impact all of the students so ... even then, course 
performance measures (tests, projects, etc.) should reflect some non 
trivial variation

of course, if the above were true, then there would be LESS variation than 
when (as is typical) students have a fairly wide range (even in upper level 
courses) of ability AND added to that impacting on variation in course 
performance measures, will be the differential impact of the course itself ...

in NO case that i can think of, no realistic case that is, would we have 
any expectation that there would be 0 variance in course performance measures

now, some might say ... well, what if we were mastery oriented in the 
course  could it not be true that at the end of the course ... everyone 
has mastered all the required skills?

the simple answer to this is NO ... if you find that all the scores at the 
end are the same ... then your measures did NOT have adequate ceiling ... 
and, you are missing detecting the more mastery that some students 
had  over others

you might think they are equal but, clearly they are not


_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Standard Deviation!

2001-11-12 Thread Dennis Roberts

i think you are asking the wrong question ... because, as far as i know ... 
there is only really one standard deviation concept ... square root of the 
variance (average of squared deviations around the mean in a set of data) ...

perhaps what you are really interested in is HOW should VARIABILITY be 
measured in your context?

perhaps the variance (ie, standard deviation) is not what you are after ...



_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: p value

2001-11-02 Thread dennis roberts

At 05:06 PM 11/2/01 -0500, Wuensch, Karl L wrote:
 Dennis wrote:   it is NOT correct to say that the p  value (as
traditionally calculated) represents the probability of finding a  result
LIKE WE FOUND  ... if the null were true? that p would be ½ of  what is
calculated.

 Jones and Tukey (A sensible formulation of the significance test,
Psychological Methods, 2000, 5, 411-414)  recently suggested that the p
which should be reported is the area of the t distribution more positive or
more negative (but not both) than the value of t obtained, just as Dennis
suggests in his post.

i would not disagree with this ... but, we have to realize that software 
(most i think) does NOT do it that way ...

if we did adopt the position of just reporting the p beyond the point you 
got ... either to the right side or left side but not both ... then, what 
will we use as the cut value ... .025??? or ... .05 as a 1 tail test? for 
rejecting the null?

we certainly will have a problem continuing to say that we set alpha at 
.05 ... in the usual two tailed sense


  ~~~
Karl L. Wuensch, Department of Psychology,
East Carolina University, Greenville NC  27858-4353
Voice:  252-328-4102 Fax:  252-328-6283
mailto:[EMAIL PROTECTED] mailto:[EMAIL PROTECTED]
http://core.ecu.edu/psyc/wuenschk/klw.htm
http://core.ecu.edu/psyc/wuenschk/klw.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



p value

2001-11-01 Thread dennis roberts

most software will compute p values (say for a typical two sample t test of 
means) by taking the obtained t test statistic ... making it both + and - 
... finding the two end tail areas in the relevant t distribution ... and 
report that as p

for example ... what if we have output like:


N  Mean StDev   SE Mean
exp   20 30.80  5.20   1.2
cont  20 27.84  3.95  0.88

Difference = mu exp - mu cont
Estimate for difference:  2.95
95% CI for difference: (-0.01, 5.92)
T-Test of difference = 0 (vs not =): T-Value = 2.02  P-Value = 0.051  DF = 35

for 35 df ... minitab finds the areas beyond -2.20 and + 2.02 ... adds them 
together .. and this value in the present case is .051

now, traditionally, we would retain the null with this p value ... and, we 
generally say that the p value means ... this is the probability of 
obtaining a result (like we got) IF the null were true

but, the result WE got was finding a mean difference in FAVOR of the exp 
group ...

however, the p value does NOT mean that the probability of finding a 
difference IN FAVOR of the exp group ... if the null were true ... is .051 
... right? since the p value has been calculated based on BOTH ends of the 
t distribution ... it includes both extremes where the exp is better than 
the control ... AND where the cont is better than the exp

thus, would it be fair to say that ... it is NOT correct to say that the p 
value (as traditionally calculated) represents the probability of finding a 
result LIKE WE FOUND  ... if the null were true? that p would be 1/2 of 
what is calculated

this brings up another point ... in the above case ... typically we would 
retain the null ... but, the p of finding the result LIKE WE DID ... if the 
null were true ... is only 1/2 of .051 ... less than the alpha of .05 that 
we have used

thus ... what alpha are we really using when we do this?

this is just a query about my continuing concern of what useful information 
p values give us ... and, if the p value provides NO (given the results we 
see) information as to the direction of the effect ... then, again ... all 
it suggests to us (as p gets smaller) is that the null is more likely  not 
to be true ...

given that it might not be true in either direction from the null ... how 
is this really helping us when we are interested in the treatment effect?

[given that we have the direction of the results AND the p value ... 
nothing else]

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Help for DL students in doing assignments

2001-10-15 Thread dennis roberts

a prime # is a natural number GREATER than 1 that can be divided ONLY by 1 
and itself ... a prime number has NO factors other than 1 and itself

i think 2 qualifies ... and is not 2 ... even?

send check to bob ASAP

At 10:40 PM 10/15/01 +0200, you wrote:

 Mr. Dawson wrote:

 Well, they do say what goes around comes around; I'd love to see 
 what
 mark the dishonest DL student gets having had his homework done for him
 by somebody who:
 
 (a) believes all primes to be odd;
 ...
###  Let's assume that any prime is NOT odd
###  It means that is is even (no other way among integers!)
###  So that prime has 3 dividers: 1,this prime and 2
###  which contradicts with prime definition:
###  (prime is integer that has only two dividers: 1 and this prime 
itself)
###  Dear Mr. Dawson, please send me at least ONE even prime
###  and i shall give you $1,000,000.

 
 -Robert Dawson






=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Are parametric assumptions importat ?

2001-10-12 Thread Dennis Roberts

At 12:59 PM 10/12/01 -0300, you wrote:

While consulting people from depts of statistics about this, a few of them
were arguing that these assumption testing are just a legend and that
there is no problem in not respecting them !

note: you should NOT respect any stat expert who says that there is no 
problem ... and not to worry about the so called classic assumptions

all they are doing is making their consultation with you EASIER for them!

every test you might want to do has 1 or more assumptions about either how 
samples were taken and/or parameters (and other things) about the population

in some cases, violations of one or more of these make little difference in 
the validity of the tests (simulation studies can verify this) ... but, 
in other cases, violations of one or more can lead to serious consequences 
(ie, yielding a much larger type I error rate for example that you thought 
you were working with) ...

there is no easy way to make some blanket statement as to what assumptions 
are important and which are not because ... this depends on a specific test 
(or family of similar tests)

usually, for a particular test ... good texts will enumerate the 
assumptions that are made AND, will give you some mini capsule of the 
impact of violations TO those assumptions




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Are parametric assumptions importat ?

2001-10-12 Thread Dennis Roberts

At 01:44 PM 10/12/01 -0400, Lise DeShea wrote:

I tell my students that the ANOVA is not robust to violation of the equal 
variances assumption, but that it's a stupid statistic anyway.  All it can 
say is either, These means are equal, or There's a difference somewhere 
among these means, but I can't tell you where it is.


i don't see that this is anymore stupid that many other null hypothesis 
tests we do ... if you want to think  stupid ... then think that it is 
stupid to think that the null can REALLY be exactly true ... so, the notion 
of doing a TEST to see if you retain or reject ... is rather stupid TOO 
since, we know that the null is NOT exactly true ... before we even do the test


_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Mean and Standard Deviation

2001-10-12 Thread dennis roberts

At 04:32 PM 10/12/01 -0500, you wrote:
A colleague of mine - not a subscriber to this helpful list - asked me if 
it is possible for the standard deviation
to be larger than the mean.  If so, under what conditions?

what about z scores??? mean = 0 and sd = 1


At first blush I do not think so  - but then I believe I have seen 
some research results in which standard deviation was larger than the mean.

Any help will be greatly appreciated..
cheersECD

___

Edward C. Dreyer
Political Science
The University of Tulsa






=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Standardized Confidence Intervals

2001-10-09 Thread dennis roberts




At 03:45 PM 10/9/01 -0400, Wuensch, Karl L wrote:
Some of those who think that estimation of the size of effects is more
important than the testing of a nil hypothesis of no effect argue that we
would be better served by reporting a confidence interval for the size of
the effect.  Such confidence intervals are, in my experience, most often
reported in terms of the original unit of measure for the variable involved.
When the unit of measure is arbitrary, those who are interested in
estimating the size of effects suggest that we do so with standardized
estimates.  It seems to me that it would be useful to present confidence
intervals in standardized units.

why? you only get further away from the original data scale/units you are 
working with ...

in what sense ... is ANY effect size indicator anything BUT arbitrary? i 
don't see how trying to standardize it ... or any confidence interval ... 
makes it anything other than still being in arbitrary units ...

i would argue that whatever the scale is you start off using ... that is as 
CLOSE as you can get to the real data ... even if the scale does not have 
any natural or intuitive kind of meaning

standardizing an arbitrary variable does NOT make it more meaningful ... 
just like converting raw data to a z score scale does NOT make the data 
more meaningful

standardizing a variable may have useful properties but, imputing more 
meaning into the raw data is not one of them



==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: Standardized Confidence Intervals

2001-10-09 Thread dennis roberts

At 03:04 PM 10/9/01 -0700, Dale Glaser wrote:

Dennis..yes, the effect size index may be arbitrary, but for argument 
sake, say I have a measure of 'self-esteem', a 10 item measure (each item 
a 5-pt. Likert scale) that has a range of 10-50;  sample1 has a 95% CI of 
[23, 27] whereas a comparison sample2 has CI of [22, 29].  Thus, by 
maintaining the CI in its own unit of measurement, we can observe that 
there is more error/wider interval for sample1 than sample2 (for now 
assuming equal 'n' for each sample).



However, it is problematic, given the inherent subjectivity of measuring 
self-esteem, to claim what is too wide of an interval for this type of 
phenomenon.

i did not know that CIs could tell you this ... under any circumstance ... 
i don't see that standardizing it will solve this problem ...

supposedly, CIs tell you something about the parameter values ... and 
nothing else ... i don't think it is within the capacity of ANY statistic 
... to tell you if some CI is too wide or too narrow ... WE have to judge 
that ... given what we consider in our heads ... is too much error or what 
we are willing to tolerate as precision of our estimates

  How do we know, especially with self-report measures, where indeed the 
 scaling may be arbitrary, if the margin of error is of concern?  It would 
 seem that by standardizing the CI, as Karl suggests, then we may be able 
 to get a better grasp of the dimensions of error...at least I know 
 the differences between .25 SD vs. 1.00 SD in terms of 
 magnitude..or is this just a stretch?!!!

you do this ahead of time ... BEFORE data are collected ... perhaps with 
some pilot work as a guide to what sds you might get ... and then you 
design it so you try to work withIN some margin of error ...

i think the underlying problem here is trying to make sense of things AFTER 
the fact ... without sufficient PREplanning to achieve some approximate 
desired result

after the fact musings will not solve what should have been dealt with 
ahead of time ... and certainly, IMHO of course, standardizing things won't 
solve this either

karl was putting regular CIs (and effect sizes) and standardized CIs (or 
effect sizes) in juxtaposition to those not liking null hypothesis testing 
but, to me, these are two different issues ...

i think that CIs and/or effect sizes are inherently more useful than ANY 
null hypothesis test ... again, IMHO ... thus, brining null hypothesis 
testing into this discussion seems not to be of value ...

of course, i suppose that debating to standardize or not standardize effect 
sizes and/or CIs ... is a legitimate matter to deal with ... even though i 
am not convinced that standardizing these things will really gain you 
anything of value

we might draw some parallel between covariance and correlation ... where, 
putting the linear relationship measure on a 'standardized' dimension IS 
useful ... so that the boundaries have some fixed limits ... which 
covariances do not ... but, i am not sure that the analog for effect sizes 
and/or CIs ... is equally beneficial


Dale N. Glaser, Ph.D.
Pacific Science  Engineering Group
6310 Greenwich Drive; Suite 200
San Diego, CA 92122
Phone: (858) 535-1661 Fax: (858) 535-1665
http://www.pacific-science.comhttp://www.pacific-science.com

-Original Message-
From: dennis roberts [mailto:[EMAIL PROTECTED]mailto:[EMAIL PROTECTED]]
Sent: Tuesday, October 09, 2001 1:52 PM
To: Wuensch, Karl L; edstat (E-mail)
Subject: Re: Standardized Confidence Intervals

dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: Standardized Confidence Intervals

2001-10-09 Thread dennis roberts

At 03:04 PM 10/9/01 -0700, Dale Glaser wrote:
  It would seem that by standardizing the CI, as Karl suggests, then we 
 may be able to get a better grasp of the dimensions of error...at 
 least I know the differences between .25 SD vs. 1.00 SD in terms of magnitude

well, yes, 1 sd means about 4 times as much spread (in sd units that is) 
than .25 sd (whether it be error or anything else) ... but, UNLess you know 
what the underlying scale is ... what the raw units mean ... have some feel 
for the metric you started with ... i don't see that this really makes it 
instantaneously more understandable

i would like to see a fully worked out example ... where we have say: 
regular effect sizes next to standardized effect sizes ... and/or regular 
CIs next to standardized CIs ... and then try to make the case that 
standardized values HELP one to UNDERSTAND the data better ... or the 
inference under examination ...

we might even do some study on this ... experimental ... where we vary the 
type of information ... then either ask Ss to elaborate on what they think 
the data mean ... or, answer some mc items about WHAT IS POSSIBLE to infer 
from the data ... and see if standardizing really makes a difference

my hunch is that it will not

..or is this just a stretch?!!!

Dale N. Glaser, Ph.D.
Pacific Science  Engineering Group
6310 Greenwich Drive; Suite 200
San Diego, CA 92122
Phone: (858) 535-1661 Fax: (858) 535-1665
http://www.pacific-science.comhttp://www.pacific-science.com


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: ranging opines about the range

2001-10-05 Thread Dennis Roberts
/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



RE: MCAS, statistics and other math problems

2001-10-05 Thread Dennis Roberts

At 07:03 AM 10/5/01 -0500, Olsen, Chris wrote:
Professor Gallagher and All --


It would appear that neither the appeal systems nor a claim of
technical adequacy would be a response to your concern about bad
questions.   The claim of technical adequacy, i.e. that good students tend
to answer them correctly anyway, but poor students don't does not, to my
mind, constitute technical adequacy.


this is absolutely correct ... all they have to go on is the score on the 
test ... if we toss in ability on something else as a defense ... then, 
why give them THIS test in the first place?




_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: MCAS, statistics and other math problems

2001-10-05 Thread dennis roberts

At 12:41 PM 10/5/01 -0500, Christopher J. Mecklin wrote:


(4) If the Massachusetts Department of Education really wants to include a 
boxplot item on the test, it should either be a multiple choice question 
written so that the correct answer is the same no matter which type of 
boxplot one was taught, or an open-ended question where the students 
actually create boxplots for 2 data sets and compare/contrast the 2 
distributions.  The readers then should be aware of both types of boxplots 
when assessing the question.

That's my two cents, anyway

actually, i think the above is worth at least 3 cents

but, the main issue re: boxplots ... is the fact that a boxplot indicates a 
median in the box, rather than a mean (say) in the box ... really 
really important ENOUGH to waste an entire question (1 in 6 about 
statistical things) on a test that is such high stakes?

seems like iF you wanted to use a boxplot as a data reporting tool ... 
within the context of an item on a test like this ... that, you would focus 
on something important like: spread of scores, or ... what is an 
approximate average value, or ... whether the distribution seems to be 
symmetrical or skewed ...


==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Help for DL students in doing assignments

2001-10-02 Thread Dennis Roberts

At 06:07 PM 10/2/01 +, Jon Miller wrote:


The neat thing about math is the numerical answer doesn't matter, just the
method.

Jon Miller

gee ... i hope you don't really mean that ... if so, that will take your 
bank off the hook IF they royally mess up your bank statement and interest 
calculations ...




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



tacoma narrows bridge

2001-10-02 Thread Dennis Roberts

At 02:36 PM 10/2/01 -0500, Olsen, Chris wrote:
Hello All --

   Not only that, I have an old Tacoma Narrows Bridge I'd like to sell
someone.


some interesting urls about this

http://www.enm.bris.ac.uk/research/nonlinear/tacoma/tacoma.html

http://www.civeng.carleton.ca/Exhibits/Tacoma_Narrows/DSmith/photos.html

http://www.civeng.carleton.ca/Exhibits/Tacoma_Narrows/ ... a little qt 
movie is here too


   -- Chris

Chris Olsen
George Washington High School
2205 Forest Drive SE
Cedar Rapids, IA

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: optimal sample size

2001-10-01 Thread Dennis Roberts

of course. the most important issue is ... what do you mean by optimal? if 
you can specify what the purpose of the sampling project is ... the 
parameter to be estimated, within what margin of error, etc. ... then you 
might be able to answer the question ... what is the MINIMAL n needed to 
accomplish these ends ... that might be optimal if you are looking for the 
smallest n you can get by with ... but, optimal does not have to be defined 
as such ...

At 12:23 PM 10/1/01 +0200, Bernhard Kuster wrote:
Hi

I am interessted in the question of optimal sample size in general, not for
a special statistical technique.

My questions: (1) What do I have to keep in mind if I compute optimal sample
size, what is relevant? (2) What are the classic studies and who has highly
influenced the subject? (3) What are the problems discussed right now by the
scientific community? (4) What are the relevant journals and is there some
information on the web?

Can anybody advise on one or more of these questions? Thanks a lot!

Bernhard



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: They look different; are they really?

2001-10-01 Thread dennis roberts

were these two different sections at the same class time? that is ... 10AM 
on mwf?
if not ... then there can be all kinds of reasons why means would be this 
different ... nonewithstanding one or two real deviant scores in either 
section ...

could also be different quality in the instruction ...

all kinds of things

of course, if you opted for 95 or 99% cis, the non overlap would be greater 
...

what is the purpose of doing this in the first place? do not the mean 
differences really suggest that there is SOMEthing different about the two 
groups ... ? or ... at least something different in the overall operation 
of the course in these two sections?

At 02:33 PM 10/1/01 -0300, Gus Gassmann wrote:
Stan Brown wrote:

  Another instructor and I gave the same exam to our sections of a
  course. Here's a summary of the results:
 
  Section A: n=20, mean=56.1, median=52.5, standard dev=20.1
  Section B: n=23  mean=73.0, median=70.0, standard dev=21.6
 
  Now, they certainly _look_ different. (If it's of any valid I can
  post the 20+23 raw data.) If I treat them as samples of two
  populations -- which I'm not at all sure is valid -- I can compute
  90% confidence intervals as follows:
 
  Class A: 48.3  mu  63.8
  Class B: 65.4  mu  80.9

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is a confidence interval?

2001-09-29 Thread dennis roberts

At 02:16 AM 9/29/01 +, John Jackson wrote:

For any random inverval selected, there is a .05% probability that the
sample will NOT yield an interval that yields the parameter being estimated
and additonally such interval will not include any values in area
represented by the left tail.  Can you make different statements about the
left and right tail?

unless CIs work differently than i think ... about 1/2 the time the CI will 
miss to the right ... and 1/2 the time they will miss to the left ... thus, 
what if we labelled EACH CI with a tag called HIT ... or MISSleft ... or 
MISSright ... for 95% CIs ... the p of grabbing a CI that is HIT from all 
possible is about .95 ... the p for getting MISSleft PLUS MISSright is 
about .05 ... thus, about 1/2 of the .05 will be MISSleft and about 1/2 of 
the .05 will be MISSright

so, i don't see that you can say anything differentially important about 
one end or the other




Michael F. [EMAIL PROTECTED] wrote in message
[EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
  (Warren) wrote in message:
 
   So, what is your best way to explain a CI?  How do you explain it
   without using some esoteric discussion of probability?
 
  I prefer to focus on the reliability of the estimate and say it is:
 
  A range of values for an estimate that reflect its unreliability and
  which contain the parameter of interest 95% of the time in the long run.




=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

======
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is a confidence interval?

2001-09-28 Thread Dennis Roberts

At 01:23 AM 9/28/01 +, Radford Neal wrote:


radford makes a nice quick summary of the basic differences between 
bayesian and frequentist positions, which is helpful. these distinctions 
are important IF one is seriously studying statistical ideas

personally, i think that trying to make these distinction for introductory 
students however is a waste of time ... these are things for majors in 
statistics or statisticians to discuss and battle over

in reference to a CI, the critical issue is CAN it be said that ... in the 
long run, there is a certain probability of producing CIs (using some CI 
construction procedure) that ... contain the parameter value ... that is, 
how FREQUENTLY we expect the CIs to contain the true value ... well, yes we can

THAT is the important idea and, i think that if we try (for the sake of 
edification of the intro student)to defend it or reject it according to 
being proper bayesian/frequentist or improper ... is totally irrelevant to 
the basic concept

but, that is just my opinion



In article yyPs7.55095$[EMAIL PROTECTED],
John Jackson [EMAIL PROTECTED] wrote:

 this is the second time I have seen this word used: frequentist? What does
 it mean?

It's the philosophy of statistics that holds that probability can
meaningfully be applied only to repeatable phenomena, and that the
meaning of a probability is the frequency with which something happens
in the long run, when the phenomenon is repeated.  This rules out
using probability to describe uncertainty about a parameter value,
such as the mass of the hydrogen atom, since there's just one true
value for the parameter, not a sequence of values.

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: E as a % of a standard deviation

2001-09-28 Thread Dennis Roberts

this is the typical margin of error formula for building a confidence 
interval were the sample mean is desired to be within a certain distance of 
the population mean

n = sample size
z = z score from nd that will produce desired confidence level (usually 
1.96 for 95% CI)
e = margin of error

so, typical CI for mu would be:

samp mean +/- z times standard error of mean

  e or the margin of error here is z * stan error of the mean (let me 
symbolize se)

 e = z * se

for 95% CI .. e = 1.96 * se

e = 1.96 * (sigma / sqrt n)

now, what n might it take to produce some e? we can rearrange the formula ...

sqrt n = (1.96 * sigma) / e

but, we don't want sqrt n ... we WANT n!

n = ((1.96 * sigma)/ e) ^2

so, what if we wanted to be within 3 points of mu with our sample mean the 
population standard deviation or sigma were 15?

   n = ((1.96 * 5) / 3)^2 = about 11 ...

only would take a SRS of about 11 to be within 3 points of the true mu 
value in your 95% confidence interval

unless i made a mistake someplace


At 09:54 AM 9/28/01 -0400, Randy Poe wrote:
John Jackson wrote:

  the forumla I was using was n = (Z?/e)^2  and attempting to express .05 
 as a
  fraction of a std dev.

I think you posted that before, and it's still getting
garbled. We see a Z followed by a question mark, and
have no idea what was actually intended.

  - Randy


=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Help with Minitab Problem?

2001-09-28 Thread dennis roberts

unless you have a million rows ... seems like using the data window and 
just sliding over each row and highlight and delete ... would be easy

by the way, why do you want to get rid of entire rows just because 
(perhaps) one value is missing? are you not wasting alot of useful data?

At 06:16 PM 9/28/01 -0400, John Spitzer wrote:
I have a dataset which has about 35 column.  Many of the cells have
missing values.  Since MINITAB recognizes the missing values, I can
perform the statistical work I need to do and don't need to worry about
the missing values.  However, I would like to be able to obtain the
subset of observations which MINITAB used for its calculations.  I would
like to be able to create a worksheet with only the rows from my dataset
which do NOT contain any missing values.

I cannot seem to find any way to do this (except manually, YUCK). Does
anyone have any ideas? suggestions?

Here is an greatly simplified version of my problem:

   C1   C2 C3  C4
1 1   2  1  2
2 2   *  *  3
3 3   3  4  2
4  0  1  *  5
5  1  *   1 8
6  2  2   3 7


I am looking for a command, or set of commands which will provide me
with the subset of the above data consisting of:

   C1   C2 C3  C4
1 1   2  1  2
2 3   3  4  2
3 2  2   3 7

Many thanks, in advance





=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



CIs

2001-09-27 Thread Dennis Roberts

it seems to me that the notion of a confidence interval is a general 
concept ... having to do with estimating some unknown quantity in which 
errors are known to occur or be present in that estimation process

in general, the generic version of a CI is:

  statistic/estimator +/- (multiplier) * error

the multiplier will drive the amount of confidence and, error will be 
estimated by varying processes depending upon the parameter or thing you 
are estimating

what we might want to estimate in a regression setting is what one 
particular person might do on a future outcome variable, like college gpa, 
given that we know what THAT person has achieved on some current variable 
(high school gpa) ... if we are interested in this specific person, then 
error will be estimated by some function of HIS/HER variation and that will 
be factored into the above generic equation as error ... this would be what 
jon cryer rightfully called a prediction interval ... BUT, it still fits 
within the realm of the concept of a CI

in other regression cases, we might not be interested in estimation for one 
specific individual on the criterion given that individual's score on the 
current variable, but rather what is the expected MEAN criterion value for 
a group of people who all got the same current variable value ... in this 
case, error is estimated by some function of the group on the current 
variable ... and this is what in regression terms is called a confidence 
band or interval ... but, the concept itself is no different than the 
prediction interval ... what IS different is what is considered error and 
how we estimate it

when we use a sample mean to estimate some population mean, we have the 
same identical general problem ... since we use the sample mean as the 
estimator and, we have a way of conceptualizing and estimating error 
(sampling error of the mean) in that case BUT, we still use the generic 
formula above ... to build our CI

in all of these cases, there is a concept of what error is and, some method 
by which we estimate it and, in all these cases we use some given quantity 
(statistic/estimator) to take a stab at an unknown quantity (parameter/true 
criterion)  and we use the estimated error around the known quantity as 
a fudge factor, tolerance factor, a margin of error factor ... when making 
our estimate of the unknown quantity of interest

all of these represent the same basic idea ... only the details of what is 
used as the point estimate and what is used as the estimate of ERROR of the 
point estimate ... change

also, in all of these cases whether it be in regression work or  sampling 
error (of means for example) work ... we still attach a quantity ... a 
percentage value ... to the intervals like  we have created when estimating 
the unknown and, as far as i can tell, we interpret that percentage in the 
same identical way in all of these cases ... with respect to the long run 
average number or percentage of hits that our intervals have of capturing 
the true value (parameter  or true criterion value)

i am more than willing to use different terms to differentiate amongst 
these different settings  ... such as in regression when you are inferring 
something about an individual ... or a group of individuals (though even 
here, i think we could select better differentiators than we currently use 
... like personal interval versus group interval) ... but overall, all of 
these are variations of the same notion and fundamental idea

IMHO of course





_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



p value

2001-09-27 Thread Dennis Roberts

let's say that you do a simple (well executed) 2 group study ... 
treatment/control ... and, are interested in the mean difference ... and 
find that a simple t test shows a p value (with mean in favor of treatment) 
of .009

while it generally seems to be held that such a p value would suggest that 
our null model is not likely to be correct (ie, some other alternative 
model might make more sense), does it say ANYthing more than that?

specifically, does the p value in and of itself impute ANY information 
about the non null possibilities being in the direction favoring the 
treatment group?

or, just that the null model is not very plausible

bottom line: is there any value added information imparted from the p value 
other than a statement about the null?

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is a confidence interval?

2001-09-27 Thread Dennis Roberts

At 07:33 AM 9/27/01 -0700, Warren wrote:


Now, we take our sample mean and s.d. and we compute a CI.  We know
we can't say anything about a probability for this single CI...it
either
contains the mean or it doesn't.  So, what DOES a CI tell us?  Does it
really give you a range of values where you think the parameter is?


most disciplines have various models for how they describe/explain/predict 
events and behaviors

in each of these, there are assumptions ... i don't see how we can get 
around that ... one must start from SOME point of reference (of course, 
some models make us start from much more stringent starting points than 
others)

to me, in statistics, particularly of the inferential type, the biggest 
assumption that we make that is suspect is the one of SRS ... taking random 
samples ...

however, if we did NOT make some assumption about the data 
being  representative of the overall population ... which SRSing helps to 
insure ... what can we do? what inferences could we possibly make?

in the case of CIs ... no, you are not sure at all that the range you got 
in your CI encompasses the parameter but, what are the odds that it does 
NOT? generally, fairly small. (well, all bets are off if you like to build 
25% CIs!) so, under these conditions, is it not reasonably assured that the 
parameter IS inside there someplace? this does not pinpoint WHERE within it 
is but, it does tend to eliminate from the long number line on which the CI 
rests ... what values do NOT seem to be too feasible for the parameter

unfortunately, if you are interested in knowing something about some 
parameter and, have no way to identify each and every population element 
and measure it (of course, even then, how do you know that your measure 
is pure?)... you are necessarily left with making this inference based on 
the data you have ... i don't see any way out of this bind ... OTHER than 
trying as best possible to take a good sample ... of decent size (to reduce 
sampling error) ... and then trusting the results that you find

if there is another way, i would certainly like to know it






_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is a confidence interval?

2001-09-26 Thread Dennis Roberts

as a start, you could relate everyday examples where the notion of CI seems 
to make sense

A. you observe a friend in terms of his/her lateness when planning to meet 
you somewhere ... over time, you take 'samples' of late values ... in a 
sense you have means ... and then you form a rubric like ... for sam ... if 
we plan on meeting at noon ... you can expect him at noon + or - 10 minutes 
... you won't always be right but, maybe about 95% of the time you will?

B. from real estate ads in a community, looking at sunday newspapers, you 
find that several samples of average house prices for a 3 bedroom, 2 bath 
place are certain values ... so, again, this is like have a bunch of means 
... then, if someone asks you (visitor) about average prices of a bedroom, 
2 bath house ... you might say ... 134,000 +/- 21,000 ... of course, you 
won't always be right but  perhaps about 95% of the time?

but, more specifically, there are a number of things you can do

1. students certainly have to know something about sampling error ... and 
the notion of a sampling distribution

2. they have to realize that when taking a sample, say using the sample 
mean, that the mean they get could fall anywhere within that sampling 
distribution

3. if we know something about #1 AND, we have a sample mean ... then, #1 
sets sort of a limit on how far away the truth can be GIVEN that sample 
mean or statistic ...

4. thus, we use the statistics (ie, sample mean) and add and subtract some 
error (based on #1) ... in such a way that we will be correct (in saying 
that the parameter will fall within the CI) some % of the time ... say, 95%?

it is easy to show this via simulation ... minitab for example can help you 
do this

here is an example ... let's say we are taking samples of size 100 from a 
population of SAT M scores ... where we assume the mu is 500 and sigma is 
100 ... i will take a 1000 SRS samples ... and summarize the results of 
building 100 CIs

MTB  rand 1000 c1-c100;  made 1000 rows ... and 100 columns ... each 
ROW will be a sample
SUBC norm 500 100.  sampled from population with mu = 500 and sigma = 100
MTB  rmean c1-c100 c101  got means for 1000 samples and put in c101
MTB  name c1='sampmean'
MTB  let c102=c101-2*10   found lower point of 95% CI
MTB  let c103=c101+2*10   found upper point of 95% CI
MTB  name c102='lowerpt' c103='upperpt'
MTB  let c104=(c102 lt 500) and (c103 gt 500)   this evaluates if the 
intervals capture 500 or not
MTB  sum c104

Sum of C104

Sum of C104 = 954.00954 of the 1000 intervals captured 500
MTB  let k1=954/1000
MTB  prin k1

Data Display

K10.954000   pretty close to 95%
MTB  prin c102 c103 c104   a few of the 1000 intervals are shown below

Data Display


  Row   lowerpt   upperpt   C104

1   477.365   517.365  1
2   500.448   540.448  0   here is one that missed 500 ...the 
other 9 captured 500
3   480.304   520.304  1
4   480.457   520.457  1
5   485.006   525.006  1
6   479.585   519.585  1
7   480.382   520.382  1
8   481.189   521.189  1
9   486.166   526.166  1
   10   494.388   534.388  1





_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: Analysis of covariance

2001-09-26 Thread Dennis Roberts

At 02:26 PM 9/26/01 -0500, Burke Johnson wrote:
 From my understanding, there are three popular ways to analyze the 
 following design (let's call it the pretest-posttest control-group design):

R Pretest   Treatment   Posttest
R PretestControl   Posttest

if random assignment has occurred ... then, we assume and we had better 
find that the means on the pretest are close to being the same ... if we 
don't, then we wonder about random assignment (which creates a mess)

anyway, i digress ...

what i would do is to do a simple t test on the difference in posttest 
means and, if you find something here ... then that means that treatment 
changed differentially compared to control

if that happens, why do anything more complicated? has not the answer to 
your main question been found?

now, what if you don't ... then, maybe something a bit more complex is 
appropriate

IMHO

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: E as a % of a standard deviation

2001-09-26 Thread Dennis Roberts

At 04:49 PM 9/26/01 +, John Jackson wrote:
re: the formula:

   n   = (Z?/e)2


could you express E as a  % of a standard deviation .

In other words does a .02 error translate into .02/1 standard deviations,
assuming you are dealing w/a normal distribution?


well, let's see ... e is the margin of error ... using the formula for a CI 
for a population mean ..

   X bar +/- z * stan error of the mean

so, the margin of error or e ... is z * standard error of the mean

now, let's assume that we stick to 95% CIs ... so the z will be about 2 ... 
that leaves us with the standard error of the mean ... or, sigma / sqrt n

let's say that we were estimating SAT M scores and assumed a sigma of about 
100 and were taking a sample size of n=100 (to make my figuring simple) ... 
this would give us a standard error of 100/10 = 10 so, the margin of error 
would be:

   e = 2 * 10 or about 20

so, 20/100 = .2 ... that is, the e or margin of error is about .2 of the 
population sd

if we had used a sample size of 400 ... then the standard error would have 
been: 100/20 = 5

and our e or margin of error would be 2 * 5 = 10

so, the margin of error is now 10/100 or .1 of a sigma unit OR 1/2 the size 
it was before

but, i don't see what you have accomplished by doing this ... rather than 
just reporting the margin of error ... 10 versus 20 ... which is also 1/2 
the size

since z * stan error is really score UNITS ... and, the way you done it ... 
.2 or .1 would represent fractions of sigma ... which still amounts to 
score UNITS ... i don't think anything new has been done ... certainly, no 
new information has been created







=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

_
dennis roberts, educational psychology, penn state university
208 cedar, AC 8148632401, mailto:[EMAIL PROTECTED]
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



Re: What is a confidence interval?

2001-09-26 Thread dennis roberts

some people are sure picky ...

given the context in which the original post was made ... it seems like the 
audience that the poster was hoping to be able to talk to about CIs was not 
very likely to understand them very well ... thus, it is not unreasonable 
to proffer examples to get one into having some sense of the notion

the examples below ... were only meant to portray ... the idea that 
observations have error ... and, over time and over samples ... one gets 
some idea about what the size of that error might be ... thus, when 
projecting about behavior ... we have a tool to know a bit about some 
penultimate value ... say the parameter for a person ... by using AN 
observation, and factoring in the error you have observed over time or 
samples ...

in essence, CIs are + and - around some observation where ... you 
conjecture within some range what the truth might be ... and, if you have 
evidence about size of error ... then, these CIs can say something about 
the parameter (again, within some range) in face of only seeing a limited 
sample of behavior

At 09:30 PM 9/26/01 +, Radford Neal wrote:
In article [EMAIL PROTECTED],
Dennis Roberts [EMAIL PROTECTED] wrote:

 as a start, you could relate everyday examples where the notion of CI seems
 to make sense
 
 A. you observe a friend in terms of his/her lateness when planning to meet
 you somewhere ... over time, you take 'samples' of late values ... in a
 sense you have means ... and then you form a rubric like ... for sam ... if
 we plan on meeting at noon ... you can expect him at noon + or - 10 minutes
 ... you won't always be right but, maybe about 95% of the time you will?
 
 B. from real estate ads in a community, looking at sunday newspapers, you
 find that several samples of average house prices for a 3 bedroom, 2 bath
 place are certain values ... so, again, this is like have a bunch of means
 ... then, if someone asks you (visitor) about average prices of a bedroom,
 2 bath house ... you might say ... 134,000 +/- 21,000 ... of course, you
 won't always be right but  perhaps about 95% of the time?

These examples are NOT analogous to confidence intervals.  In both
examples, a distribution of values is inferred from a sample, and
based on this distribution, a PROBABILITY statement is made concerning
a future observation.  But a confidence interval is NOT a probability
statement concerning the unknown parameter.  In the frequentist
statistical framework in which confidence intervals exists,
probability statements about unknown parameters are not considered to
be meaningful.

you are clearly misinterpreting, for whatever purpose, what i have said

i certainly have NOT said that a CI is a probability statement about any 
specific parameter or, being able to attach some probability value to some 
certain value as BEING the parameter

the p or confidence associated with CIs only makes sense in terms of 
dumping all possible CIs into a hat ... and, asking  what is the 
probability of pulling one out at random that captures the parameter 
(whatever the parameter might be) ...

the example i gave with some minitab work clearly showed that ... and made 
no other interpretation about p values in connection with CIs

perhaps some of you who seem to object so much to things i offer ... might 
offer some posts of your own in response to requests from those seeking 
help ... to make sure that they get the right message ...


Radford Neal


Radford M. Neal   [EMAIL PROTECTED]
Dept. of Statistics and Dept. of Computer Science [EMAIL PROTECTED]
University of Toronto http://www.cs.utoronto.ca/~radford



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
   http://jse.stat.ncsu.edu/
=

==
dennis roberts, penn state university
educational psychology, 8148632401
http://roberts.ed.psu.edu/users/droberts/drober~1.htm



=
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
  http://jse.stat.ncsu.edu/
=



  1   2   3   4   5   >