This is a multi-part message in MIME format.
------=_NextPart_000_0044_01C267BB.33A51F80
Content-Type: text/plain;
charset="iso-8859-1"
Content-Transfer-Encoding: quoted-printable
Gus,
You seem to be playing the game "heads Gus wins, tails Bill loses." =
There is nothing dishonest about what I am saying. I describe a =
procedure that is very easy to do with any spread sheet. The logic is =
there to support the results. When you say "admit" perhaps you should =
say that you see. No one is guilty here and no one is winning. We are =
scientists seeking truth, together. So you SEE that we can differetiate =
the cause from the effect, assuming that effects are combinations of =
causes... at least in simulations and analyses as I describe them and as =
you have replicated them. What you are seeing is what many people before =
you have seen and gone on to dismiss by reference to what they believe =
is my crazy personality or some other nonsensical excuse for backing out =
when the results support CR. I trust that you will not do that but I do =
not know if any demonstration or logic is capable of convencing you of =
the validity of CR. Can you be convinced or are you simply trying to =
disprove me with out any concern for the truth? I do not mean this =
offensively, but I do want to have some sense of where this conversation =
is headed.=20
I am afraid that I still do not know what you did. Remember, I am only =
a psychologist, not a mathematician. Would you mind giving us a small =
data set and step by step instructions, for example, as we might do with =
excel? I am particularly concerned about the criteria you use to pick =
sets of values. I do not see the point in your sampling scheme since =
all we need to do is sort the variables by the normally distributed ones =
and trim off all the data corresponding to the upper and lower tails of =
the normal distributions. This seems to be the most parsimonious thing =
to do. Furthermore, when the normally distributed variable is the =
hypthesized effect, we need not ever trim the tails off. The effect is =
supposed to be triangular.=20
It sounds to me that when you collect subsamples you are selecting y =
values somehow so that you are building in additional dependencies =
between the collected x values and the y values. This is why I need to =
know exactly how you select the values to retain.
I am sure that had you failed to replicate the trimming experiment as I =
described it you would have claimed the failure to be proof - that CR =
does not measure causation. Heads Gus wins, tails Bill loses. Lets be =
explicit. I will win no matter what happens to CR because I will better =
understand the truth. Everything both of us are doing is loaded. We are =
putting many assumptions and discoveries on the line by doing tests. It =
is the sort of thing real men do. But let us be explicit.
So define your terms, philosopher, and let the axe fall where is does.=20
Best,
Bill
"Gus Gassmann" <[EMAIL PROTECTED]> wrote in message =
[EMAIL PROTECTED]">news:[EMAIL PROTECTED]...
=20
[EMAIL PROTECTED] wrote:=20
Gus,=20
I am still not sure what you are doing. What is a bucket? The =
essence of=20
what you seem to be claiming is that when we sample y to be uniform, =
then CR=20
gives us the opposite results. You admit, however, that CR works =
with the=20
usual approach.
"Works" is a loaded word. I admit that when x1 and x2 are sampled to =
be uniform=20
and you compute y =3D x1 + x2 and further compute the correlation of y =
with x1=20
both overall and over subsets of the x1 (lower, middle and upper), the =
correlation=20
coefficients follow a predicted pattern.=20
Now here is another stab at explaining what I did. First off, when you =
generate=20
x1 uniformly on [a1,b1] and x2 uniformly on [a2,b2], then you can set =
up=20
a grid on the rectangle [a1,b1] x [a2,b2], like this: (Hope this comes =
out;=20
it is intended to be viewed with a monospaced font.)=20
(a1,b2) (b1,b2)=20
---- ---- ---- ----=20
| | | | | |=20
| | | | ... | |=20
---- ---- ---- ----=20
| | | | | |=20
| | | | | |=20
---- ---- ---- ----=20
=20
...=20
=20
---- ---- ---- ----=20
| | | | | |=20
| | | | ... | |=20
---- ---- ---- ----=20
(a1,a2) (b1,a2)=20
If the distribution is uniform, you would expect each of the smaller=20
rectangles to contain roughly the same number of points (provided=20
the sample is large enough). My idea was to force this relationship=20
by _picking_ one set of values in each tiny rectangle. If you have=20
10000 tiny rectangles (100 x 100), that gives you a pretty good=20
approximation to a uniform distribution (with 10000 data points).=20
(Do you agree?)=20
So then I set out to find a subsample of the big one that was uniform=20
in x2 and y (where y was computed earlier as y =3D x1 + x2). Here's =
the=20
important question: Do you agree that the causality in the subsample=20
should be the same as the causality in the overall sample?=20
To my surprise and amazement, I did not find the pattern I predicted.=20
In fact, what this smaller sample suggests is that x2 and y cause x1.=20
In other words, different selection of the data points results in a=20
different causal relationship. This should not happen.=20
------=_NextPart_000_0044_01C267BB.33A51F80
Content-Type: text/html;
charset="iso-8859-1"
Content-Transfer-Encoding: quoted-printable
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD>
<META http-equiv=3DContent-Type content=3D"text/html; =
charset=3Diso-8859-1">
<META content=3D"MSHTML 5.50.4522.1800" name=3DGENERATOR>
<STYLE></STYLE>
</HEAD>
<BODY bgColor=3D#ffffff>
<DIV><FONT face=3DArial size=3D2>Gus,</FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>You seem to be playing the game "heads =
Gus wins,=20
tails Bill loses." There is nothing dishonest about =
what I am=20
saying. I describe a procedure that is very easy to do with =
any=20
spread sheet. The logic is there to support the results. When you say =
"admit"=20
perhaps you should say that you see. No one is guilty here and no =
one=20
is winning. We are scientists seeking truth, together. So you SEE=20
that we can differetiate the cause from the effect, assuming that =
effects=20
are combinations of causes... at least in simulations and analyses as I =
describe=20
them and as you have replicated them. What you are seeing is what many =
people=20
before you have seen and gone on to dismiss by reference to what they =
believe is=20
my crazy personality or some other nonsensical excuse for backing out =
when the=20
results support CR. I trust that you will not do that but I do not know =
if any=20
demonstration or logic is capable of convencing you of the validity of =
CR. =20
Can you be convinced or are you simply trying to disprove me with out =
any=20
concern for the truth? I do not mean this offensively, but I do =
want to=20
have some sense of where this conversation is headed. </FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>I am afraid that I still do not know =
what you=20
did. Remember, I am only a psychologist, not a mathematician. =
Would you=20
mind giving us a small data set and step by step instructions, for =
example, as=20
we might do with excel? I am particularly concerned about the criteria =
you use=20
to pick sets of values. I do not see the point in your sampling =
scheme=20
since all we need to do is sort the variables by the normally =
distributed ones=20
and trim off all the data corresponding to the upper and lower tails of =
the=20
normal distributions. This seems to be the most parsimonious =
thing=20
to do. Furthermore, when the normally distributed variable is the=20
hypthesized effect, we need not ever trim the tails off. The effect is =
supposed=20
to be triangular. </FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT><FONT face=3DArial =
size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>It sounds to me that when you collect =
subsamples=20
you are selecting y values somehow so that you are building in =
additional=20
dependencies between the collected x values and the y values. This is =
why I need=20
to know exactly how you select the values to retain.</FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>I am sure that had you failed to =
replicate the=20
trimming experiment as I described it you would have claimed the failure =
to be=20
proof - that CR does not measure causation. Heads Gus wins, tails =
Bill=20
loses. Lets be explicit. I will win no matter what happens to CR =
because I=20
will better understand the truth. Everything both of us are doing is=20
loaded. We are putting many assumptions and discoveries on the =
line by=20
doing tests. It is the sort of thing real men do. But let us be=20
explicit.</FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>So define your terms, philosopher, and =
let the axe=20
fall where is does. </FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>Best,</FONT></DIV>
<DIV><FONT face=3DArial size=3D2></FONT> </DIV>
<DIV><FONT face=3DArial size=3D2>Bill</FONT></DIV>
<BLOCKQUOTE dir=3Dltr=20
style=3D"PADDING-RIGHT: 0px; PADDING-LEFT: 5px; MARGIN-LEFT: 5px; =
BORDER-LEFT: #000000 2px solid; MARGIN-RIGHT: 0px">
<DIV>"Gus Gassmann" <<A=20
href=3D"mailto:[EMAIL PROTECTED]">[EMAIL PROTECTED]</A>> =
wrote in=20
message <A=20
=
href=3D"[EMAIL PROTECTED]"">news:[EMAIL PROTECTED]">news:[EMAIL PROTECTED]=
al.ca</A>...</DIV> =20
<P><A href=3D"mailto:[EMAIL PROTECTED]">[EMAIL PROTECTED]</A> =
wrote:=20
<BLOCKQUOTE TYPE=3D"CITE">Gus,=20
<P>I am still not sure what you are doing. What is a bucket? =
The=20
essence of <BR>what you seem to be claiming is that when we sample y =
to be=20
uniform, then CR <BR>gives us the opposite results. You admit, =
however, that CR works with the <BR>usual =
approach.</P></BLOCKQUOTE>"Works" is=20
a loaded word. I admit that when x1 and x2 are sampled to be uniform =
<BR>and=20
you compute y =3D x1 + x2 and further compute the correlation of y =
with x1=20
<BR>both overall and over subsets of the x1 (lower, middle and upper), =
the=20
correlation <BR>coefficients follow a predicted pattern.=20
<P>Now here is another stab at explaining what I did. First off, when =
you=20
generate <BR>x1 uniformly on [a1,b1] and x2 uniformly on [a2,b2], then =
you can=20
set up <BR>a grid on the rectangle [a1,b1] x [a2,b2], like this: (Hope =
this=20
comes out; <BR>it is intended to be viewed with a monospaced font.)=20
=
<P>(a1,b2) &nb=
sp; &nbs=
p;  =
; =20
(b1,b2) <BR><TT> ---- ----=20
---- ----</TT> =
<BR><TT> =20
| | | =20
| | |</TT>=20
<BR><TT> | | =20
| | ... | |</TT>=20
<BR><TT> ---- ----=20
---- ----</TT> =
<BR><TT> =20
| | | =20
| | |</TT>=20
<BR><TT> | | =20
| | | =
|</TT>=20
<BR><TT> ---- ----=20
---- ----</TT> <BR><TT> </TT> =
<BR><TT> ...</TT> <BR> =20
<BR><TT> ---- ----=20
---- ----</TT> =
<BR><TT> =20
| | | =20
| | |</TT>=20
<BR><TT> | | =20
| | ... | |</TT>=20
<BR><TT> ---- ----=20
---- ----</TT>=20
=
<BR>(a1,a2) &n=
bsp; &nb=
sp; &nbs=
p; =20
(b1,a2)=20
<P>If the distribution is uniform, you would expect each of the =
smaller=20
<BR>rectangles to contain roughly the same number of points (provided =
<BR>the=20
sample is large enough). My idea was to force this relationship <BR>by =
_picking_ one set of values in each tiny rectangle. If you have =
<BR>10000 tiny=20
rectangles (100 x 100), that gives you a pretty good <BR>approximation =
to a=20
uniform distribution (with 10000 data points). <BR>(Do you agree?)=20
<P>So then I set out to find a subsample of the big one that was =
uniform=20
<BR>in x2 and y (where y was computed earlier as y =3D x1 + x2). =
Here's the=20
<BR>important question: Do you agree that the causality in the =
subsample=20
<BR>should be the same as the causality in the overall sample?=20
<P>To my surprise and amazement, I did not find the pattern I =
predicted.=20
<BR>In fact, what this smaller sample suggests is that x2 and y cause =
x1.=20
<BR>In other words, different selection of the data points results in =
a=20
<BR>different causal relationship. This should not happen.=20
</P></BLOCKQUOTE></BODY></HTML>
------=_NextPart_000_0044_01C267BB.33A51F80--
.
.
=================================================================
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at:
. http://jse.stat.ncsu.edu/ .
=================================================================