Russ, Nick,

You both make an interesting point about one of these dimensions (unpredictability) requiring an observer, while the other (organization, or structure) does not.

However, Heinz von Foerster, I believe, would disagree. I believe he would say that BOTH require an observer!

Another way to think about this is: Science is by definition an empirical enterprise. It requires dispassionate observation, refutable observations (Popper), etc. Science therefore requires an observer. The very idea of components being "related", or not, into an "organization" or "structure" is itself an abstraction on the part of an observer.

Of course this thread of thought takes us away from our main point of discussion about entropy...which I will continue on my next missive. ;-)

Grant

Russ Abbott wrote:
If you call it behavioral rather than predictable it doesn't require a predictor. It's just an arrangement in time.

-- Russ



On Sat, Aug 7, 2010 at 12:14 PM, Nicholas Thompson <[email protected] <mailto:[email protected]>> wrote:

    Grant –

Glad you are on board, here. I will read this carefully. Does this have anything to do with the Realism Idealism thing. Predictibility requires a person to be predicting; organization is
    there even if there is no one there to predict one part from another.

    N

    *From:* [email protected]
    <mailto:[email protected]>
    [mailto:[email protected]
    <mailto:[email protected]>] *On Behalf Of *Grant Holland
    *Sent:* Saturday, August 07, 2010 2:06 PM
    *To:* [email protected]; The Friday Morning Applied Complexity
    Coffee Group

    *Subject:* Re: [FRIAM] entropy and uncertainty, REDUX

    Russ - Yes.



    I use the terms "organizational" and "predictable", rather than
    "structural" and "behavioral", because of my particular interests.
    They amount to the same ideas. Basically they are two orthogonal
    dimensions of certain state spaces as they change.

    I lament the fact that the same term "entropy" is used to apply to
    both meanings, however. Especially since few realize that these
    two meanings are being conflated with the same word. Von Foerster
    actually defined the word "entropy" in two different places within
    the same book of essays to mean each of these two meanings! Often
    the word "disorder" is used. And people don't know whether
    "disorder" refers to "disorganization" or whether it refers to
    "unpredictability". This word has fostered the further unfortunate
    confusion.

    It seems few people make the distinction that you have. This
    conflation causes no end of confusion. I really wish there were 2
    distinct terms. In my work, I have come up with the acronym
    "DOUPBT" for the "unpredictable" meaning of entropy. (Or,
    "behavioral", as you call it.) This stands for Degree Of
    UnPredictaBiliTy.) I actually use Shannon's formula for this meaning.

    This all came about because 1) Clausius invented the term entropy
    to mean "dissipation" (a kind of dis-organization, in my terms).
    2) But then Gibbs came along and started measuring the degree of
    unpredictability involved in knowing the "arrangements" (positions
    and momenta) of molecules in an ideal gas. The linguistic problem
    was that Gibbs (and Boltzmann) use the same term - entropy - as
    had Clausius, even though Clausius emphasized a structural
    (dissipation) idea, whereas Gibbs emphasized an unpredictability
    idea (admittedly, unpredictability of "structural" change).

    To confuse things even more, Shannon came along and defined
    entropy in purely probabilistic terms - as a direct measure of
    unpredictability. So, historically, the term went from a purely
    structural meaning, to a mixture of structure and unpredictability
    to a pure unpredictability meaning. No wonder everyone is confused.

    Another matter is that Clausius, Boltzmann and Gibbs were all
    doing Physics. But Shannon was doing Mathematics.

    My theory is Mathematics. I'm not doing Physics. So I strictly
    need Shannon's meaning. My "social problem" is that every time I
    say "entropy", too many people assume I'm talking about
    "dissipation" when I am not. I'm always talking about
    "disorganization" when I use the term in my work. So, I have gone
    to using the phrase "Shannon's entropy", and never the word in its
    naked form. (Admittedly, I eventually also combine in a way
    similar to Gibbs :-[ . But I do not refer to the combined result
    as "entropy".)

    :-P
    Grant


    Russ Abbott wrote:

    Is it fair to say that Grant is talking about what one might call
    structural vs. behavioral entropy?

    Let's say I have a number of bits in a row. That has very low
    structural entropy. It takes very few bits to describe that row of
    bits. But let's say each is hooked up to a random signal. So
    behaviorally the whole thing has high entropy. But the behavioral
    uncertainty of the bits is based on the assumed randomness of the
    signal generator. So it isn't really the bits themselves that have
    high behavioral entropy. They are just a "window" through which we
are observing the high entropy randomness behind them.
    This is a very contrived example. Is it at all useful for a
    discussion of structural entropy vs. behavioral entropy? I'm
    asking that in all seriousness; I don't have a good sense of how
    to think about this.

    This suggests another thought. A system may have high entropy in
    one dimension and low entropy in another. Then what? Most of us
    are very close to the ground most of the time. But we don't stay
    in one place in that relatively 2-dimensional world. This sounds a
    bit like Nick's example. If you know that an animal is female, you
    can predict more about how she will act than if you don't know that.

    One other thought Nick talked about gradients and the tendency for
    them to dissipate.  Is that really so? If you put two mutually
    insoluble liquids in a bottle , one heavier than another, the
    result will be a layer cake of liquids with a very sharp gradient
    between them. Will that ever dissipate?

    What I think is more to the point is that potential energy
    gradients will dissipate. Nature abhors a potential energy
    gradient -- but not all gradients.


    -- Russ

    On Thu, Aug 5, 2010 at 11:09 AM, Grant Holland
    <[email protected] <mailto:[email protected]>>
    wrote:

    Glen is very close to interpreting what I mean to say. Thanks, Glen!

    (But of course, I have to try one more time, since I've  thought
    of another - hopefully more compact - way to approach it...)

    Logically speaking, "degree of unpredictability" and "degree of
    disorganization" are orthogonal concepts and ought to be able to
    vary independently - at least in certain domains. If one were to
    develop a theory about them (and I am), then that theory should
    provide for them to be able to vary independently.

    Of course, for some "applications" of that theory, these
    "predictability/unpredictability" and
    "organization/disorganization" variables may be dependent on each
    other. For example, in Thermodynamics, it may be that the degree
    unpredictability and the degree of disorganization are correlated.
    (This is how many people seem to interpret the second law.) But
    this is specific to a Physics application.

    However, in other applications, it could be that the degree
    uncertainty and the degree of disorganization vary independently.
    For example, I'm developing a mathematic theory of living and
    lifelike systems. Sometimes in that domain there is a high degree
    of predictability that an organo-chemical entity is organized, and
    sometimes there is unpredictability around that. The same
    statement goes for predictability or unpredictability around
disorganization. Thus, in the world of living systems, unpredictability and disorganization can vary independently.

    To make matters more interesting, these two variables can be
    joined in a joint space. For example, in the "living systems
    example" we could ask about the probability of advancing from a
    certain disorganized state in one moment to a certain organized
    state in the next moment. In fact, we could look at the entire
    probability distribution of advancing from this certain
    disorganized state at this moment to all possible states at the
    next moment - some of which are more disorganized than others. But
    if we ask this question, then we are asking about a probability
    distribution of states that have varying degrees of organization
    associated with them. But, we also have a probability distribution
    involved now, so we can ask "what is it's Shannon entropy?" That
    is, what is its degree of unpredictability? So we have created a
    joint space that asks about both disorganization and
    unpredictability at the same time. This is what I do in my theory
    ("Organic Complex Systems").

    Statistical Thermodynamics (statistical mechanics) also mixes
    these two orthogonal variables in a similar way. This is another
    way of looking at what Gibbs (and Boltzmann) contributed.
    Especially Gibbs talks about the probability distributions of
    various "arrangements" (organizations) of molecules in an ideal
    gas (these arrangements, states, are defined by position and
    momentum). So he is interested in probabilities of various
    "organizations" of molecules. And, the Gibbs formula for entropy
    is a measurement of this combination of interests. I suspect that
    it is this combination that is confusing to so many. (Does
    "disorder" mean "disorganization", or does it mean
    "unpredictability". In fact, I believe reasonable to say that
    Gibbs formula measures "the unpredictability of being able to talk
    about which "arrangements" will obtain."

    In fact, Gibbs formula for thermodynamic entropy looks exactly
    like Shannon's - except for the presence of a constant in Gibbs
    formula. They are isomorphic! However, they are speaking to
    different domains. Gibbs is modeling a physics phenomena, and
    Shannon is modeling a mathematical statistics phenomena. The
    second law applies to Gibbs conversation - but not to Shannon's.

    In my theory, I use Shannon's - but not Gibbs'.

    (Oops, I guess that wasn't any shorter than Glen's explanation. :-[ )

    Grant



    glen e. p. ropella wrote:

    Nicholas Thompson wrote  circa 08/05/2010 08:30 AM:

        All of this, it seems to me, can be accommodated by – indeed requires –

        a common language between information entropy and physics entropy, the

        very language which GRANT seems to argue is impossible.

    OK.  But that doesn't change the sense much.  Grant seemed to be arguing

    that it's because we use a common language to talk about the two

    concepts, the concepts are erroneously conflated.  I.e. Grant not only

    admits the possibility of a common language, he _laments_ the common

    language because it facilitates the conflation of the two different

    concepts ... unless I've misinterpreted what he's said, of course.

        I would like to apologize to everybody for these errors.  I am beginning

        to think I am too old to be trusted with a distribution list.  It’s not

        that I don’t go over the posts before I send them … and in fact, what I

        sent represented weeks of thinking and a couple of evenings of drafting

        … believe it or not!  It seems that there are SOME sorts of errors I

        cannot see until they are pointed out to me, and these seem to be, of

        late, the fatal ones.

    We're all guilty of this.  It's why things like peer review and

    criticism are benevolent gifts from those who donate their time and

    effort to criticize others.  It's also why e-mail and forums are more

    powerful and useful than the discredit they usually receive.  While it's

    true that face-to-face conversation has higher bandwidth, e-mail,

    forums, and papers force us to think deeply and seriously about what we

    say ... and, therefore think.  So, as embarrassing as "errors" like this

    feel, they provide the fulcrum for clear and critical thinking.  I say

    let's keep making them!

    Err with Gusto! ;-)

--
    Grant Holland

    VP, Product Development and Software Engineering

    NuTech Solutions

    404.427.4759


    ============================================================
    FRIAM Applied Complexity Group listserv
    Meets Fridays 9a-11:30 at cafe at St. John's College
    lectures, archives, unsubscribe, maps at http://www.friam.org

    ------------------------------------------------------------------------

    ============================================================

    FRIAM Applied Complexity Group listserv

    Meets Fridays 9a-11:30 at cafe at St. John's College

    lectures, archives, unsubscribe, maps at http://www.friam.org



--
    Grant Holland

    VP, Product Development and Software Engineering

    NuTech Solutions

    404.427.4759


    ============================================================
    FRIAM Applied Complexity Group listserv
    Meets Fridays 9a-11:30 at cafe at St. John's College
    lectures, archives, unsubscribe, maps at http://www.friam.org


------------------------------------------------------------------------

============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

--
Grant Holland
VP, Product Development and Software Engineering
NuTech Solutions
404.427.4759

============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

Reply via email to