[EMAIL PROTECTED] wrote:
> Paul and Gus,
>
> The point of finding a subsample is to get rid of the normal distribution
> and replace it with a uniform sample that allows the concurrence of similar
> extremes of both causes. If we start with a normal distribution and want a
> uniform, then a random subsample will NOT do. All that would do is give us
> back a picture of the same original normal sample, but with smaller sample
> size.
>
> I can not seem to communicate that normal distributions are not better than
> others, just because we often get normal distributions from snap shots.
> Just going out and taking a snap shot of some phenomenon is likely to leave
> us with the extremes of the causes under sampled... a normal distribution.
> I know this statement challenges a very large number of studies in many
> disciplines. But I was not the first to so challenge. Nunnally called
> normally sampled causes "dead data." Only the effects should be sampled as
> normal or triangular.
My ignorance shows, since I have no clue as to who Nunnally was.
At any rate, I think he is quite wrong. My problem with his position
is that it _presupposes_ that I know what are the causes and the
effects. If I then want to use CR to classify my variables, that is
circular logic.
Here are two data sets for you:
Sample 1 Sample 2
x1 x2 x3 x1 x2 x3
-0.1080 0.3953 0.5033 -0.4397 0.2960 -0.1437
-0.0397 -0.4221 -0.3824 0.1654 -0.0392 0.1262
0.1731 -0.3578 -0.5309 -0.1806 -0.3556 -0.5362
-0.1922 -0.3609 -0.1687 0.3486 -0.2583 0.0903
-0.4954 0.0686 0.5640 -0.3397 -0.2943 -0.6340
0.1204 0.2726 0.1522 0.2225 0.2985 0.5210
0.2326 -0.1835 -0.4161 0.4383 0.1177 0.5560
0.0382 -0.1677 -0.2059 0.4894 -0.1643 0.3251
0.1101 -0.2875 -0.3976 0.2245 0.3840 0.6085
0.2314 -0.0712 -0.3026 -0.1449 -0.1220 -0.2669
0.1834 -0.0154 -0.1988 0.3106 -0.3532 -0.0426
-0.3144 -0.3424 -0.0280 -0.0671 -0.2752 -0.3423
-0.1609 -0.1717 -0.0108 -0.0526 0.1395 0.0869
-0.3071 -0.0672 0.2399 -0.0737 -0.4318 -0.5055
0.3023 0.3298 0.0275 -0.4424 -0.1317 -0.5741
-0.3086 -0.1721 0.1365 -0.2790 -0.3178 -0.5968
-0.0301 0.4549 0.4850 0.1514 0.2256 0.3770
-0.4170 0.4049 0.8219 0.3591 -0.1085 0.2506
0.3822 0.4298 0.0476 0.0932 -0.3287 -0.2355
0.4678 -0.3920 -0.8598 -0.2843 0.3962 0.1119
-0.3454 0.4602 0.8056 -0.1186 0.0616 -0.0570
-0.1997 -0.4107 -0.2110 0.3543 0.4832 0.8375
-0.0753 -0.2161 -0.1408 -0.0579 -0.3364 -0.3943
0.4514 0.3021 -0.1493 0.0820 0.4140 0.4960
-0.4914 0.1282 0.6196 0.0238 0.2010 0.2248
-0.3188 -0.2597 0.0591 -0.4660 -0.0083 -0.4743
0.0954 -0.4192 -0.5146 0.3160 0.1450 0.4610
0.3109 -0.1859 -0.4968 0.2707 -0.3065 -0.0358
0.1177 0.0425 -0.0752 -0.1334 -0.4000 -0.5334
0.0988 0.0866 -0.0122 0.3131 0.0778 0.3909
0.2042 -0.3173 -0.5215 0.1473 -0.3525 -0.2052
0.4936 0.0075 -0.4861 0.1552 0.3760 0.5312
0.3995 0.1239 -0.2756 -0.4542 0.3861 -0.0681
-0.2928 0.1044 0.3972 -0.1404 0.3967 0.2563
-0.2079 -0.4417 -0.2338 0.0219 -0.1143 -0.0924
-0.2872 0.2315 0.5187 -0.2758 0.0582 -0.2176
0.0952 0.2241 0.1289 0.1723 0.1686 0.3409
0.0612 0.4615 0.4003 0.0937 -0.4729 -0.3792
0.0994 -0.3605 -0.4599 -0.0574 0.0202 -0.0372
-0.2700 -0.0366 0.2334 0.0330 0.1393 0.1723
0.2352 -0.2092 -0.4444 -0.1287 0.4526 0.3239
-0.3092 0.3306 0.6398 0.4807 -0.4881 -0.0074
-0.3007 0.2191 0.5198 0.1418 0.0555 0.1973
-0.0645 -0.1181 -0.0536 -0.1412 0.2659 0.1247
0.1709 -0.1709 -0.3418 0.0594 -0.2961 -0.2367
0.4761 0.4055 -0.0706 -0.4084 -0.3958 -0.8042
-0.4904 0.2683 0.7587 -0.3724 0.4771 0.1047
-0.2681 0.0731 0.3412 -0.0559 -0.0793 -0.1352
-0.0226 0.2264 0.2490 -0.0836 0.2387 0.1551
0.1847 0.1867 0.0020 0.2396 0.0202 0.2598
0.2076 0.2120 0.0044 -0.3015 0.0792 -0.2223
0.2451 0.4303 0.1852 -0.3462 0.3331 -0.0131
-0.0490 -0.0084 0.0406 -0.1550 -0.0568 -0.2118
-0.0365 -0.3069 -0.2704 -0.3262 -0.3329 -0.6591
0.2611 0.3574 0.0963 0.4767 0.2800 0.7567
-0.0611 0.1078 0.1689 -0.1874 -0.2461 -0.4335
-0.1591 -0.0597 0.0994 0.0916 0.0904 0.1820
0.3689 -0.2308 -0.5997 -0.4325 0.0419 -0.3906
0.2228 -0.4240 -0.6468 -0.4178 -0.4326 -0.8504
0.4775 -0.2383 -0.7158 0.4032 -0.2920 0.1112
0.3783 -0.0090 -0.3873 0.3121 0.2828 0.5949
-0.0566 0.0260 0.0826 0.3388 0.3410 0.6798
0.4026 0.1592 -0.2434 0.0499 0.3937 0.4436
-0.1845 0.1619 0.3464 -0.2070 -0.2163 -0.4233
-0.2490 -0.1894 0.0596 0.1267 0.4613 0.5880
0.3668 -0.3443 -0.7111 -0.0371 0.4682 0.4311
-0.4651 -0.0808 0.3843 -0.0616 0.3536 0.2920
0.3750 0.0377 -0.3373 0.2678 -0.0612 0.2066
-0.1645 -0.2539 -0.0894 0.0980 -0.0522 0.0458
0.4041 -0.1345 -0.5386 -0.2904 -0.0675 -0.3579
0.0645 -0.2520 -0.3165 0.1373 -0.1961 -0.0588
0.0621 0.3327 0.2706 -0.1248 0.1215 -0.0033
-0.1591 0.4404 0.5995 -0.3287 -0.4614 -0.7901
0.1182 -0.4835 -0.6017 -0.3457 0.2191 -0.1266
-0.2415 -0.2480 -0.0065 0.2812 0.4961 0.7773
-0.3003 -0.4429 -0.1426 0.1968 -0.2689 -0.0721
0.1662 0.3290 0.1628 -0.4027 0.4364 0.0337
-0.2538 0.4813 0.7351 0.4085 0.3003 0.7088
0.2485 0.1853 -0.0632 -0.3424 -0.1506 -0.4930
-0.2859 -0.3326 -0.0467 -0.2163 0.4987 0.2824
-0.1710 0.0792 0.2502 -0.2549 0.2514 -0.0035
0.0548 -0.0992 -0.1540 -0.0244 -0.1422 -0.1666
0.0410 0.1996 0.1586 0.4855 0.4630 0.9485
0.3219 0.2388 -0.0831 0.3642 -0.0608 0.3034
0.4776 0.2809 -0.1967 -0.3695 -0.0575 -0.4270
-0.0407 0.3851 0.4258 0.2506 -0.2512 -0.0006
-0.3385 0.0563 0.3948 -0.2974 -0.4193 -0.7167
0.3362 -0.4555 -0.7917 -0.4036 -0.2230 -0.6266
-0.4863 -0.3058 0.1805 0.2822 -0.1218 0.1604
-0.3244 0.1590 0.4834 -0.2181 0.1220 -0.0961
-0.2982 0.3868 0.6850 -0.4685 0.1107 -0.3578
-0.4723 -0.1808 0.2915 0.4669 0.0894 0.5563
-0.4189 -0.2992 0.1197 -0.3594 0.1894 -0.1700
0.1484 0.4720 0.3236 0.4702 -0.3301 0.1401
0.4847 -0.4153 -0.9000 0.4001 -0.0267 0.3734
-0.4401 0.3119 0.7520 0.2147 -0.4939 -0.2792
0.2471 0.0177 -0.2294 0.3135 -0.4004 -0.0869
-0.1915 0.2655 0.4570 0.2022 0.1466 0.3488
-0.4083 -0.4815 -0.0732 0.1384 -0.4058 -0.2674
0.4440 -0.0507 -0.4947 -0.2001 -0.1060 -0.3061
In both samples the columns marked x1 and x2 are uniformly distributed
on [-0.5, +0.5]. (Reasonably closely, anyway: the data are set up so that
there will be one (x1,x2) in [-0.5,-0.4]x[-0.5,-0.4], one in
[-0.5,-0.4]x[-0.4,-0.3],
one in [-0.5,-0.4]x[-0.3,-0.2], and so on, for the obvious partition of
[-0.5,+0.5]x[-0.5,+0.5] into 100 smaller squares of side 0.1. So they do
fill the square, as you requested.)
Both, in fact are subsets of larger samples, and I can
send you the full files if interested --- there is not enough room here.
In both samples the three columns are connected via x3 = x1 + x2.
However, in one of the two samples x3 is in fact computed in this way,
caused by the other two, the way I use the word, whereas in the other
sample x1 and x3 are originally generated (uniformly) and then x2
is calculated as x2 = x3 - x1, in other words, x2 is the _effect_.
Can you tell them apart?
> Sampling is a type of perception and like all perception, the instrument is
> a construction. The eyes do not allow us to see many wave lengths. We are
> subject to certain illusions. The eyes are not perfect. No sampling strategy
> is perfect either. Only God in his omniscience has perfect perception. We
> ain't him. Neither is the normal distribution.
> When we collect an approximately equal number AND a sufficient number of
> observations so that we can see the combinations of all the possible values
> of the causes, then we get a less distorted view of how those causes can
> combine. The manifold is a map or prototype of such structure. A cause can
> be expressed in data but unless we can see it, we can not make much sense of
> it. This is why scientists search long and hard to fill out the ranges of
> their observations in experimental science. Normal distributions do not
> reflect the combinations of the similar extremes of causes. You do not see
> normally distributed cell sizes across the factors of an ANOVA. We need
> something else to see what happens in the extremes of the
> cross-tabulation/conjunction of causes. This something else is a manifold,
> which is approximated by uniform samples of the causes.
>
> Bill
.
.
=================================================================
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at:
. http://jse.stat.ncsu.edu/ .
=================================================================