Re: Disadvantage of Non-parametric vs. Parametric Test

1999-12-07 Thread Glen Barnett

Alex Yu wrote:
 
 Disadvantages of non-parametric tests:
 
 Losing precision: Edgington (1995) asserted that when more precise
 measurements are available, it is unwise to degrade the precision by
 transforming the measurements into ranked data.

So this is an argument against rank-based nonparametric tests
rather than nonparametric tests in general. In fact, I think
you'll find Edgington highly supportive of randomization procedures,
which are nonparametric.

In fact, surprising as it may seem, a lot of the location 
information in a two sample problem is in the ranks. Where
you really start to lose information is in ignoring ordering
when it is present.
 
 Low power: Generally speaking, the statistical power of non-parametric
 tests are lower than that of their parametric counterpart except on a few
 occasions (Hodges  Lehmann, 1956; Tanizaki, 1997).

When the parametric assumptions hold, yes. e.g. if you assume normality
and the data really *are* normal. When the parametric assumptions are
violated, it isn't hard to beat the standard parametric techniques.

However, frequently that loss is remarkably small when the parametric
assumption holds exactly. In cases where they both do badly, the
parametric may outperform the nonparametric by a more substantial
margin (that is, when you should use something else anyway - for
example, a t-test outperforms a WMW when the distributions are
uniform).

 Inaccuracy in multiple violations: Non-parametric tests tend to produce
 biased results when multiple assumptions are violated (Glass, 1996;
 Zimmerman, 1998).

Sometimes you only need one violation:
Some nonparametric procedures are even more badly affected by
some forms of non-independence than their parametric equivalents.
 
 Testing distributions only: Further, non-parametric tests are criticized
 for being incapable of answering the focused question. For example, the
 WMW procedure tests whether the two distributions are different in some
 way but does not show how they differ in mean, variance, or shape. Based
 on this limitation, Johnson (1995) preferred robust procedures and data
 transformation to non-parametric tests.

But since WMW is completely insensitive to a change in spread without
a change in location, if either were possible, a rejection would 
imply that there was indeed a location difference of some kind. This
objection strikes me as strange indeed. Does Johnson not understand
what WMW is doing? Why on earth does he think that a t-test suffers
any less from these problems than WMW?
 
Similarly, a change in shape sufficient to get a rejection of a WMW
test would imply a change in location (in the sense that the "middle"
had moved, though the term 'location' becomes somewhat harder to pin
down precisely in this case).  e.g. (use a monospaced font to see this):

:. .:
::.   =  .::
...   ...
a b   a b
 
would imply a different 'location' in some sense, which WMW will
pick up. I don't understand the problem - a t-test will also reject
in this case; it suffers from this drawback as well (i.e. they are
*both* tests that are sensitive to location differences, insensitive
to spread differences without a corresponding location change, and
both pick up a shape change that moves the "middle" of the data).

However, if such a change in shape were anticipated, simply testing
for a location difference (whether by t-test or not) would be silly. 

Nonparametric (notably rank-based) tests do have some problems,
but making progress on understanding just what they are is 
difficult when such seemingly spurious objections are thrown in.

His preference for robust procedures makes some sense, but the
preference for (presumably monotonic) transformation I would 
see as an argument for a rank-based procedure. e.g. lets say
we are in a two-sample situation, and we decide to use a t-test
after taking logs, because the data are then reasonably normal...
in that situation, the WMW procedure gives the same p-value as 
for the untransformed data. However, let's assume that the 
log-transform wasn't quite right... maybe not strong enough. When 
you finally find the "right" transformation to normality, there
you finally get an extra 5% (roughly) efficiency over the WMW you
started with. Except of course, you never know you have the right
transformation - and if the distribution the data are from are
still skewed/heavy-tailed after transformation (maybe they were
log-gamma to begin with or something), then you still may be better
off using WMW.

Do you have a full reference for Johnson? I'd like to read what
the reference actually says.

Glen



UNSUBSCRIBE UNSUBSCRIBE

1999-12-07 Thread Jose Rojo



   \\|//
   (o o)
+--oOO--(_)--OOo-+
| Jose Manuel Rojo   | tlf  : +++ 91 + 5-64-29-63|
| C.S.I.C  , SPAIN   | Email: [EMAIL PROTECTED]|
++ 
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-==-=-=-==-=-=-=-
o   \ o /   _o __|\ / |__ o _   \ o /  o
   /|\| /\   __\o\o|o/ o/__   /\  |   /|\   
   / \   / \   | \   /) |( \  /o\  / ) |  (\  / |/ \  / \   
=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-



Re: Scale Reliability

1999-12-07 Thread John Donovan

Brett -

Although you may have a conceptual reason for summing the four item scores to
form a composite, the low level of internal consistency that these items display
suggests that they may not be assessing a unitary construct.  So, to sum these
four items and label this score as a single variable or construct (locus of
control?) may be problematic.  Have you assessed the dimensionality of these
items?  If you find that these items do not strongly load on a single,  unitary
factor, you may want to consider breaking the items up, rather than summing
them.

John





Re: Scale Reliability

1999-12-07 Thread PChawla



The fact that the shorter scale has low internal consistency doesn't necessarily
mean that the 4 items in question are not unidimensional.  It may just be that
the measurement error is large relative to their covariance.  Given that the
four items in question are drawn from a scale with established internal
consistency, I'd suspect they probably are measuring the same thing - only not
measuring it very well.

purnima.




To:
cc:
From: John Donovan [EMAIL PROTECTED]/
Date:  12/07/99 01:09:40 PM

Subject:  Re: Scale Reliability



Brett -

Although you may have a conceptual reason for summing the four item scores to
form a composite, the low level of internal consistency that these items display
suggests that they may not be assessing a unitary construct.  So, to sum these
four items and label this score as a single variable or construct (locus of
control?) may be problematic.  Have you assessed the dimensionality of these
items?  If you find that these items do not strongly load on a single,  unitary
factor, you may want to consider breaking the items up, rather than summing
them.

John











Re: stats packages for Unix

1999-12-07 Thread David Cross/Psych Dept/TCU

This is great ... I had lost my reference to this site and product.
Thanks!



On Sat, 4 Dec 1999, KO wrote:

 Hello:
 
 I would suggest R (aka GNU-S). You cane
 find more information about it at
 http://www.stat.cmu.edu/R/CRAN/
 
 I highly recommend it.
 
 Sincerely,
 
 Kouros Owzar ([EMAIL PROTECTED])
 
 
 
 
 Bob Hayden [EMAIL PROTECTED] wrote in message
 [EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
  Any advice on basic stats. packages for Unix/Linux?  We used Minitab
  on a DEC alpha box using DEC Unix.  That was fine by me but the DEC
  was too expensive to maintain.  So, we replaced it with a FreeBSD
  system runing on a pile of Intel boxes.  But, Minitab does not run on
  that, so we started using the PC and iMac versions.  However, the iMac
  lab is a disaster (local problem, not a Mac problem, I think) and we
  hate supporting three platforms.
 
  The software must be VERY easy to use and not cost much more than
  Minitab, which is quite cheap for a site license.  We have a few
  hundred students per year using the software.  Most of them have
  pretty limited math., computer and study skills.
 
 
_
   | |  Robert W. Hayden
   | |  Department of Mathematics
  /  |  Plymouth State College MSC#29
 |   |  Plymouth, New Hampshire 03264  USA
 | * |  Rural Route 1, Box 10
/|  Ashland, NH 03217-9702
   | )  (603) 968-9914 (home)
   L_/  [EMAIL PROTECTED]
fax (603) 535-2943 (work)
 
 
 



Re: Software for logistic regression

1999-12-07 Thread David Cross/Psych Dept/TCU

SYSTAT has an excellent log regression procedure and it may be cheap
enough for you.

On Sun, 5 Dec 1999, FourCubed wrote:

 Hello everyone!  Can anyone tell me the least expensive software for performing
 a (conditional) logistic regression.  SPSS and SAS are too expensive, and I
 don't have access to a university computer with existing software.  Thank you.
 
 Steve
 [EMAIL PROTECTED]
 



Re: Help for my dissertation

1999-12-07 Thread J. Williams

First, check to see if your school doesn't have a stat lab available.  Most 
major universities have a site where graduate students and others involved in 
empirical research can get free assistance.  Often, it is in the form of 
anothe, but more advanced, graduate student, but many of these folks are 
perfectly capable of looking over your proposal and can give helpful hints.  
Second, you might want to consider adding a statistically knowledgeable person 
to your committee if it is not too late.  If I were you, I would get the 
statistical guidance BEFORE you start collecting data or whatever methodology 
you envision.  Lastly, you can go to the statistics department at your school 
and get the names of consultants available for looking over your materials.
Good Luck
j. williams

In article 82j1a2$6s3$[EMAIL PROTECTED], [EMAIL PROTECTED] wrote:
I am a 3rd year student carrying out resaerch for
my dissertation. I require help with stats -
primarily to confirm that I amusing the correct
methods to test my hypothesis and secondly to
understand them once I have collected my data.  I
am prepared to pay for some ones time if they
could help me.  Is there anyone out there that
could help?


Sent via Deja.com http://www.deja.com/
Before you buy.



STATISTICA group

1999-12-07 Thread Dr. Wolfgang Hitzl

Dear STATISTICA users,

in order to subscribe to the  mailing list/group for STATISTICA
users, visit

  http://www.computer.privateweb.at/Statistica/

and  type just in your e-mail adress and follow the instructions.
Then you will be added to our list.

My very best wishes

Dr. Wolfgang Hitzl
(group moderator)






SDT: What is D-sub-A and Sakitt's D?

1999-12-07 Thread Victor A. Gombos

[Please forgive me if this esoteric question was answered before in this
newsgroup]

I am using Systat 9.0 for my master's thesis data--the nature of my
analyses depend heavily on Signal Detection Theory. Therefore, of course

I am using the Signal Detection Analysis program in Systat.

It seems to me that using d' is not as good as D-sub-a or Sakitt's D --
the subjects' responses I am analyzing are confidence ratings, on a
scale from one-to-eight -- as there tends to be greater
fluctuations/variability of d-prime as opposed to the other measures.
In this way,  I'm committed to D-sub-a thinking it is more robust and
consistent.

But I need to know how D-sub-A and Sakitt's D differ from d'.  I haven't

been able to find  information on this from any other source thus far.

Can anyone tell me what these measures are and how they differ from d'?

Thanks,
V.



Re: Help for my dissertation

1999-12-07 Thread SAlbert

I'd second the comments of finding local help; that's definitely the best
option.  If, for whatever reason, that's not available, then, and only then,
consider looking for someone at a distance to help out.  If you do have to do
that, offer more detail about what kind of study you're doing:  the type of
data, how it's collected, how it's analyzed, what questions you want to
investigate, etc.; that's all necessary before anyone (remote or local) can
determine what you need and whether they can help you.

Steve Albert