On 10/9/24 12:33 PM, glen wrote:
Hm. I don't normally read the GPT output posted to the list. But I did
this time and am worse off for it. Your original is way better.
Anyway, I'd like to continue responding to Stephen's originalism and
your tangent into compression at the same time.
Because wrong or wrong-headed, or is that a false dichotomy?
The idea that models are abstract is what I think Jon was targeting
with the bisimulation reference. There's no requirement that models be
less detailed than their referents. In fact, I'd argue that many
models have more details than their referent (ignoring von Neumann's
interpretation of Goedel for a minute). A mathematical model of, say,
a toy airplane in a wind tunnel (especially such a model implemented
in code - with the whole stack from programming language down to
transistors) feels more packed with detail than the actual toy
airplane in the wind tunnel. There's detail upon detail in that model.
Were we to take assembly theory seriously, I think the implemented
model is way more detailed (complicated) than the referent. What makes
the model useful isn't the compression, but the determinism ... the
rational/mental control over the mechanism. We have less control over
the toy airplane and the wind than we have over the tech stack in
which the model is implemented.
I'm not sure what it means in this context for the model to be more
detailed than the referent? The toy airplane in a wind tunnel is likely
less detailed (or differently detailed?) than a real airplane in a real
airflow but if the (computer?) model of such is *more detailed* then I
think this invokes your "excess meaning" or "excess detail?" dismissal?
if the computer model/simulation has more detail than the
toy-airplane/wind-tunnel model of the fully elaborated "real" airplane
in "real airflow" then does it have *more* than the latter? If so, not
only excess but also "wrong"? If more detailed than the toy/wind-tunnel
then simply closer to referent?
Here's where aphorisms and "models" writ large differ drastically.
Aphorisms are purposely designed to have Barnum-Forer (heuristic)
power. Models often have that (especially in video games, movies,
etc.) power.
I appreciate this distinction/acknowledgment.
But Good Faith modelers work against that. A Good Faith modeler will
pepper their models' uses with screaming large print IT'S JUST A MODEL.
I agree there is virtue in acknowledging the implications of "IT"S JUST
A MODEL!!!!"
Aphorisms (and all psuedo-profound bullshit) are not used that way.
And are all aphorisms by definition "pseudo-profound bullshit"? Or do
they still retain some profoundish-utility? do they in any way represent
a (finessed?) useful compression?
They are used in much the same way *inductive* models are used to
trick you into thinking the inference from your big data is
categorically credible.
I agree with Stephen that Box is mostly referring to inductive
inference. But then again, with the demonstrative power of really
large inductively tuned models, we're starting to blur the lines
between induction, deduction, and abduction. That false trichotomy was
profound at some point. But these days, sticking to it like Gospel is
problematic.
I accept de-rigeur that "sticking to anything like Gospel" is
problematic. I'm a little slow at the switch here on the earlier part
of the paragraph.. I will study it. It reads at least mildly profound
and I trust not "pseudo-so".
On 10/9/24 10:49, steve smith wrote:
Now the original for the 0 or 2 people who might have endured this far:
The first clause (protasis?) seems to specifically invoke the
"dimension reduction" implications of "compression" but some of the
recent discussion here seems to invoke the "discretization" or more
aptly perhaps the "limited precision"? I think the stuff about
bisimulation is based on this difference?
The trigger for this flurry of "arguing about words" was Wilson's:
"We have Paleolithic emotions, medieval institutions, and
god-like technology."
to which there were varioius objections ranging from (paraphrasing):
"it is just wrong"
"this has been debunked"
to the ad-hominem:
"Wilson was once good at X but he should not be listened to
for Y"
The general uproar *against* this specific aphorism seemed to be
a proxy for:
"it is wrong-headed" and "aphorisms are wrong-headed" ?
then Glen's objection (meat on the bones of "aphorisms are
wrong-headed"?) that aphorisms are "too short" which is what lead me
to thinking about aphorisms as models, models as a form or expression
of compression and the types of compression (lossy/not) and how that
might reflect the "bisimulation" concept
https://en.wikipedia.org/wiki/Bisimulation . At first I had the
"gotcha" or "aha" response to learning more about bisimulation that
it applied exclusively/implicitly to finite-state systems but in fact
it seems that as long as there is an abstraction that obscures or
avoids any "precision" issues it applies to all state-transition
systems.
This lead me to think about the two types of compression that
models (or aphorisms?) offer. One breakdown of the features of
compression in modeling are: Abstraction; Dimension Reduction; Loss
of Detail; Pattern Recognition. The first and last (abstraction
and pattern recognition) seem to be features/goals of modeling, The
middle two seem to be utilitarian while the loss of detail is more of
a bug, an inconvenience nobody values (beyond the utility of keeping
the model small and in the way it facilitates "pattern recognition"
in a ?perverse? way)
-. --- - / ...- .- .-.. .. -.. / -- --- .-. ... . / -.-. --- -.. .
FRIAM Applied Complexity Group listserv
Fridays 9a-12p Friday St. Johns Cafe / Thursdays 9a-12p Zoom
https://bit.ly/virtualfriam
to (un)subscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/
archives: 5/2017 thru present https://redfish.com/pipermail/friam_redfish.com/
1/2003 thru 6/2021 http://friam.383.s1.nabble.com/