Grant

 

You see that I am still struggling to understand your original distinction
between prediction and organization.  Clearly organization affords
prediction.   I was trying to read you as saying that organization is the
thing that's there and prediction is what we make of it.  We can use
organizations to make predictions.  But you just blocked that
interpretation.  

 

I apologize if I am not reading astutely enough.  If you or somebody else
could help me out, here, I would be in your debt.  What is the distinction
that you see between these two things that seem so much the same to me. 

 

Nick 

 

From: [email protected] [mailto:[email protected]] On Behalf
Of Grant Holland
Sent: Saturday, August 07, 2010 3:25 PM
To: [email protected]; The Friday Morning Applied Complexity Coffee
Group
Subject: Re: [FRIAM] entropy and uncertainty, REDUX

 

Russ, Nick,

You both make an interesting point about one of these dimensions
(unpredictability) requiring an observer, while the other (organization, or
structure) does not.

However, Heinz von Foerster, I believe, would disagree. I believe he would
say that BOTH require an observer!

Another way to think about this is: Science is by definition an empirical
enterprise. It requires dispassionate observation, refutable observations
(Popper), etc. Science therefore requires an observer. The very idea of
components being "related", or not, into an "organization" or "structure" is
itself an abstraction on the part of an observer.

Of course this thread of thought takes us away from our main point of
discussion about entropy...which I will continue on my next missive. ;-) 

Grant

Russ Abbott wrote: 

If you call it behavioral rather than predictable it doesn't require a
predictor. It's just an arrangement in time.



-- Russ





On Sat, Aug 7, 2010 at 12:14 PM, Nicholas Thompson
<[email protected]> wrote:

Grant -

 

Glad you are on board, here.  I will read this carefully.  

 

Does this have anything to do with the Realism Idealism thing.
Predictibility requires a person to be predicting; organization is there
even if there is no one there to predict one part from another. 

 

N

 

From: [email protected] [mailto:[email protected]] On Behalf
Of Grant Holland
Sent: Saturday, August 07, 2010 2:06 PM
To: [email protected]; The Friday Morning Applied Complexity Coffee
Group 


Subject: Re: [FRIAM] entropy and uncertainty, REDUX

 

Russ - Yes.



I use the terms "organizational" and "predictable", rather than "structural"
and "behavioral", because of my particular interests. They amount to the
same ideas. Basically they are two orthogonal dimensions of certain state
spaces as they change.

I lament the fact that the same term "entropy" is used to apply to both
meanings, however. Especially since few realize that these two meanings are
being conflated with the same word. Von Foerster actually defined the word
"entropy" in two different places within the same book of essays to mean
each of these two meanings! Often the word "disorder" is used. And people
don't know whether "disorder" refers to "disorganization" or whether it
refers to "unpredictability". This word has fostered the further unfortunate
confusion. 

It seems few people make the distinction that you have. This conflation
causes no end of confusion. I really wish there were 2 distinct terms. In my
work, I have come up with the acronym "DOUPBT" for the "unpredictable"
meaning of entropy. (Or, "behavioral", as you call it.) This stands for
Degree Of UnPredictaBiliTy.) I actually use Shannon's formula for this
meaning.

This all came about because 1) Clausius invented the term entropy to mean
"dissipation" (a kind of dis-organization, in my terms). 2) But then Gibbs
came along and started measuring the degree of unpredictability involved in
knowing the "arrangements" (positions and momenta) of molecules in an ideal
gas. The linguistic problem was that Gibbs (and Boltzmann) use the same term
- entropy - as had Clausius, even though Clausius emphasized a structural
(dissipation) idea, whereas Gibbs emphasized an unpredictability idea
(admittedly, unpredictability of "structural" change).

To confuse things even more, Shannon came along and defined entropy in
purely probabilistic terms - as a direct measure of unpredictability. So,
historically, the term went from a purely structural meaning, to a mixture
of structure and unpredictability to a pure unpredictability meaning. No
wonder everyone is confused.

Another matter is that Clausius, Boltzmann and Gibbs were all doing Physics.
But Shannon was doing Mathematics. 

My theory is Mathematics. I'm not doing Physics. So I strictly need
Shannon's meaning. My "social problem" is that every time I say "entropy",
too many people assume I'm talking about "dissipation" when I am not. I'm
always talking about "disorganization" when I use the term in my work. So, I
have gone to using the phrase "Shannon's entropy", and never the word in its
naked form. (Admittedly, I eventually also combine in a way similar to Gibbs
:-[ . But I do not refer to the combined result as "entropy".)

:-P 
Grant


Russ Abbott wrote: 

Is it fair to say that Grant is talking about what one might call structural
vs. behavioral entropy?

Let's say I have a number of bits in a row. That has very low structural
entropy. It takes very few bits to describe that row of bits. But let's say
each is hooked up to a random signal. So behaviorally the whole thing has
high entropy. But the behavioral uncertainty of the bits is based on the
assumed randomness of the signal generator. So it isn't really the bits
themselves that have high behavioral entropy. They are just a "window"
through which we are observing the high entropy randomness behind them.  

This is a very contrived example. Is it at all useful for a discussion of
structural entropy vs. behavioral entropy? I'm asking that in all
seriousness; I don't have a good sense of how to think about this.

This suggests another thought. A system may have high entropy in one
dimension and low entropy in another. Then what? Most of us are very close
to the ground most of the time. But we don't stay in one place in that
relatively 2-dimensional world. This sounds a bit like Nick's example. If
you know that an animal is female, you can predict more about how she will
act than if you don't know that. 

One other thought Nick talked about gradients and the tendency for them to
dissipate.  Is that really so? If you put two mutually insoluble liquids in
a bottle , one heavier than another, the result will be a layer cake of
liquids with a very sharp gradient between them. Will that ever dissipate?

What I think is more to the point is that potential energy gradients will
dissipate. Nature abhors a potential energy gradient -- but not all
gradients.


-- Russ 

 

On Thu, Aug 5, 2010 at 11:09 AM, Grant Holland <[email protected]>
wrote:

Glen is very close to interpreting what I mean to say. Thanks, Glen!

(But of course, I have to try one more time, since I've  thought of another
- hopefully more compact - way to approach it...)

Logically speaking, "degree of unpredictability" and "degree of
disorganization" are orthogonal concepts and ought to be able to vary
independently - at least in certain domains. If one were to develop a theory
about them (and I am), then that theory should provide for them to be able
to vary independently. 

Of course, for some "applications" of that theory, these
"predictability/unpredictability" and "organization/disorganization"
variables may be dependent on each other. For example, in Thermodynamics, it
may be that the degree unpredictability and the degree of disorganization
are correlated. (This is how many people seem to interpret the second law.)
But this is specific to a Physics application.

However, in other applications, it could be that the degree uncertainty and
the degree of disorganization vary independently. For example, I'm
developing a mathematic theory of living and lifelike systems. Sometimes in
that domain there is a high degree of predictability that an organo-chemical
entity is organized, and sometimes there is unpredictability around that.
The same statement goes for predictability or unpredictability around
disorganization.  Thus, in the world of  living systems,  unpredictability
and  disorganization can vary independently. 

To make matters more interesting, these two variables can be joined in a
joint space. For example, in the "living systems example" we could ask about
the probability of advancing from a certain disorganized state in one moment
to a certain organized state in the next moment. In fact, we could look at
the entire probability distribution of advancing from this certain
disorganized state at this moment to all possible states at the next moment
- some of which are more disorganized than others. But if we ask this
question, then we are asking about a probability distribution of states that
have varying degrees of organization associated with them. But, we also have
a probability distribution involved now, so we can ask "what is it's Shannon
entropy?" That is, what is its degree of unpredictability? So we have
created a joint space that asks about both disorganization and
unpredictability at the same time. This is what I do in my theory ("Organic
Complex Systems").

Statistical Thermodynamics (statistical mechanics) also mixes these two
orthogonal variables in a similar way. This is another way of looking at
what Gibbs (and Boltzmann) contributed. Especially Gibbs talks about the
probability distributions of various "arrangements" (organizations) of
molecules in an ideal gas (these arrangements, states, are defined by
position and momentum). So he is interested in probabilities of various
"organizations" of molecules. And, the Gibbs formula for entropy is a
measurement of this combination of interests. I suspect that it is this
combination that is confusing to so many. (Does "disorder" mean
"disorganization", or does it mean "unpredictability". In fact, I believe
reasonable to say that Gibbs formula measures "the unpredictability of being
able to talk about which "arrangements" will obtain."

In fact, Gibbs formula for thermodynamic entropy looks exactly like
Shannon's - except for the presence of a constant in Gibbs formula. They are
isomorphic! However, they are speaking to different domains. Gibbs is
modeling a physics phenomena, and Shannon is modeling a mathematical
statistics phenomena. The second law applies to Gibbs conversation - but not
to Shannon's.

In my theory, I use Shannon's - but not Gibbs'.

(Oops, I guess that wasn't any shorter than Glen's explanation. :-[ )

Grant 



glen e. p. ropella wrote: 

Nicholas Thompson wrote  circa 08/05/2010 08:30 AM:
  

All of this, it seems to me, can be accommodated by - indeed requires -
a common language between information entropy and physics entropy, the
very language which GRANT seems to argue is impossible.
    

OK.  But that doesn't change the sense much.  Grant seemed to be arguing
that it's because we use a common language to talk about the two
concepts, the concepts are erroneously conflated.  I.e. Grant not only
admits the possibility of a common language, he _laments_ the common
language because it facilitates the conflation of the two different
concepts ... unless I've misinterpreted what he's said, of course.
 
  

I would like to apologize to everybody for these errors.  I am beginning
to think I am too old to be trusted with a distribution list.  It's not
that I don't go over the posts before I send them . and in fact, what I
sent represented weeks of thinking and a couple of evenings of drafting
. believe it or not!  It seems that there are SOME sorts of errors I
cannot see until they are pointed out to me, and these seem to be, of
late, the fatal ones.
    

 
We're all guilty of this.  It's why things like peer review and
criticism are benevolent gifts from those who donate their time and
effort to criticize others.  It's also why e-mail and forums are more
powerful and useful than the discredit they usually receive.  While it's
true that face-to-face conversation has higher bandwidth, e-mail,
forums, and papers force us to think deeply and seriously about what we
say ... and, therefore think.  So, as embarrassing as "errors" like this
feel, they provide the fulcrum for clear and critical thinking.  I say
let's keep making them!
 
Err with Gusto! ;-)
 
  

 

-- 
Grant Holland
VP, Product Development and Software Engineering
NuTech Solutions
 
404.427.4759


============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

 

 

  _____  

 
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

 

-- 
Grant Holland
VP, Product Development and Software Engineering
NuTech Solutions
404.427.4759


============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

 

 

  _____  

 
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org





-- 
Grant Holland
VP, Product Development and Software Engineering
NuTech Solutions
404.427.4759
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

Reply via email to