Rich Ulrich <[EMAIL PROTECTED]> wrote in sci.stat.edu:
>[ posted and e-mailed.]
Ditto.
>On Sat, 29 Dec 2001 16:46:10 -0500, [EMAIL PROTECTED] (Stan Brown)
>wrote:
>> Now we come to the part I'm having conceptual trouble with: "Have
>> you proven that one gas gives better mileage than the other? If so,
>> which one is better?"
>>
>> Now obviously if the two are different then one is better, and if
>> one is better it's probably B since B had the higher sample mean.
>
>I want to raise an eyebrow at this earlier statement.
Hmm... Which "earlier statement" do you mean? If two means are
different then one of them _must_ be larger than the other; that's
how real numbers work. Can you explain your raised eyebrow a bit
more specifically? Or is it just the word "proven", about which I
comment below.
> We should
>not overlook the chance to teach our budding statisticians:
>*Always* pay attention to the distinction between random trials
>or careful controls, on the one hand; and grab-samples on the other.
>[Maybe your teacher asked the question that way, in order to
>lead up to that in class?]
No; this was in a book of homework problems, which is pretty
standard at the junior college where I teach. Specifically, it was a
lengthy exercise in using Excel to do the sort of statistical tests
the students normally do on a TI83.
>The numbers do not *prove* that one gas gives better mileage;
>the mileage was, indeed, better for one gas than another -- for
>reasons yet to be discussed. Different cars? drivers? routes?
All good points for discussion. But I wouldn't focus too much on
that off-the-cuff word "prove". (I'm not being defensive since I
didn't write the exercise. :-) My students did understand that
nothing is ever proved; that there's still a p-value chance of
getting the sample results you got even if you did perfect random
selection an d the null hypothesis is true. Maybe I'm being UNDER-
scrupulous here, but I think it a pardonable bit of sloppy language.
>> But are we in fact justified in jumping from a two-tailed test (=/=)
>> to a one-tailed result (>)?
>>
>> Here we have a tiny p-value, and in fact a one-tailed test gives a
>> p-value of 0.0001443. But something seems a little smarmy about
>> first setting out to discover whether there is a difference -- just
>> a difference, unequal means -- then computing a two-tailed test and
>> deciding to announce a one-tailed result.
>
>Another small issue. Why did the .00014 appear?
I added that for purposes of posting; the original exercise didn't
have the students do a one-tailed test at all. It's just half the
two-tailed p-value, as I'm sure you recognize.
>In clinical trials, we observe the difference and then we do attribute
>it to one end. But it is not the convention to report the one-tailed
>p-level, after the fact. I think there are editors who would object
>to that, but that is a guess. Also, for various reasons, our smallest
>p-level for reporting is usually 0.001.
Well, these two p-values are smaller than that: you're talking
significance level of 0.1% and these were 0.014% or 0.029%.
But my question was not about reporting a smaller p-value; it was
about first establishing a two-tailed "difference" and then moving
from that to declaring which side the difference lies on. I think
A.G. McDowell has disposed of that, however.
--
Stan Brown, Oak Road Systems, Cortland County, New York, USA
http://oakroadsystems.com/
"My theory was a perfectly good one. The facts were misleading."
-- /The Lady Vanishes/ (1938)
=================================================================
Instructions for joining and leaving this list and remarks about
the problem of INAPPROPRIATE MESSAGES are available at
http://jse.stat.ncsu.edu/
=================================================================