If you're interested, I've written a wiki
page<http://cs.calstatela.edu/wiki/index.php/Entropy>that describes
entropy. The goal was to make the concept both rigorous and
intuitive.  If you look at it, let me know where it fails.

*-- Russ *



On Mon, Jun 6, 2011 at 9:08 AM, ERIC P. CHARLES <e...@psu.edu> wrote:

> Nick,
> The notion of information that Shannon proposes takes a very idealized
> understanding of "communication." I think it is a good model for machine
> "communication" and things like that (i.e., metaphorical communication), but
> it will not make you very happy, what with your feet-on-the-ground study of
> actual communication between organisms.  For example, as I understand
> Shannon's information theory, there must be countless things transmitted
> from one organism to another that do not count as information, but which
> nevertheless are 'sent' by one organism and alter the behavior of the other.
> Also, we cannot have a conversation over whether or not it is in the
> interests of the organism to base their behavior on the information they
> receive form other organisms, because 'information' has been defined as that
> on which it is good to base behavior. Also, also, we also cannot talk about
> the transmission of information already known by the receiver, because if it
> is already known, then the message is not information. That is, if 1) we are
> flipping a coin, 2)  I see the coin land heads, 3) you say 'heads', then
> your message contained no information.
>
> Eric
>
> P.S. Oddly, for the last point, I probably need to say that your message
> contained no information 'about the coin.' In information theory land they
> don't want to count it as information that your saying 'heads' tells me that
> you also have seen the coin as landing a heads (i.e., they don't want to
> count the information it gives me about you). If they counted that, then all
> messages would contain information.
>
>
> On Mon, Jun 6, 2011 09:44 AM, *"Nicholas Thompson" <
> nickthomp...@earthlink.net>* wrote:
>
>  Grant,
>
>
>
> This seems backwards to me, but I got properly thrashed for my last few
> postings so I am putting my hat over the wall very carefully here.
>
>
>
> I thought……i thought …. the information in a message was the number of
> bits by which the arrival of the message decreased the uncertainty of the
> receiver.  So, let’s say you are sitting awaiting the result of a coin toss,
> and I am on the other end of the line flipping the coin.  Before I say
> “heads” you have 1 bit of uncertainty; afterwards, you have none.
>
>
>
> The reason I am particularly nervous about saying this is that it, of
> course, holds out the possibility of negative information.   Some forms of
> communication, appeasement gestures in animals, for instance, have the
> effect of increasing the range of behaviors likely to occur in the
> receiver.  This would seem to correspond to a negative value for the
> information calculation.
>
>
>
> Nick
>
> *From:* friam-boun...@redfish.com [mailto:friam-boun...@redfish.com] *On
> Behalf Of *Grant Holland
> *Sent:* Sunday, June 05, 2011 11:07 PM
> *To:* The Friday Morning Applied Complexity Coffee Group; Steve Smith
> *Subject:* Re: [FRIAM] Quote of the week
>
>
>
> Interesting note on "information" and "uncertainty"...
>
> Information is Uncertainty. The two words are synonyms.
>
> Shannon called it "uncertainty", contemporary Information theory calls it
> "information".
>
> It is often thought that the more information there is, the less
> uncertainty. The opposite is the case.
>
> In Information Theory (aka the mathematical theory of communications) , the
> degree of information I(E) - or uncertainty U(E) - of an event is measurable
> as an inverse function of its probability, as follows:
>
> U(E) = I(E) = log( 1/Pr(E) ) = log(1) - log( Pr(E) ) = -log( Pr(E) ).
>
> Considering I(E) as a random variable, Shannon's entropy is, in fact, the
> first moment (or expectation) of I(E). Shannon entropy = exp( I(E) ).
>
> Grant
>
> On 6/5/2011 2:20 PM, Steve Smith wrote:
>
>
>
> *"Philosophy is to physics as pornography is to sex. It's cheaper, it's
> easier and some people seem to prefer it."*
>
>
> Modern Physics is  contained in Realism which is contained in Metaphysics
> which I contained in all of Philosophy.
>
> I'd be tempted to counter:
>
> *"Physics is to Philosophy as the Missionary Position is to the Kama
> Sutra"*
>
>
> Physics also appeals to Phenomenology and Logic (the branch of Philosophy
> were Mathematics is rooted) and what we can know scientifically is
> constrained by Epistemology (the nature of knowledge) and phenomenology (the
> nature of conscious experience).
>
> It might be fair to say that many (including many of us here) who hold
> Physics up in some exalted position simply dismiss or choose to ignore all
> the messy questions considered by  *the rest of* philosophy.   Even if we
> think we have clear/simple answers to the questions, I do not accept that
> the questions are not worthy of the asking.
>
> The underlying point of the referenced podcast is, in fact, that Physics,
> or Science in general might be rather myopic and limited by it's own
> viewpoint by definition.
>
> * "The more we know, the less we understand."*
>
>
> Philosophy is about understanding, physics is about knowledge first and
> understanding only insomuch as it is a part of natural philosophy.
>
> Or at least this is how my understanding is structured around these
> matters.
>
> - Steve
>
>  On Sun, Jun 5, 2011 at 1:15 PM, Robert Holmes 
> <rob...@holmesacosta.com<#13065b895b5ce71e_>>
> wrote:
>
> >From the BBC's science podcast "The Infinite Monkey 
> >Cage<http://www.bbc.co.uk/podcasts/series/timc>
> ":
>
> "Philosophy is to physics as pornography is to sex. It's cheaper, it's
> easier and some people seem to prefer it."
>
>
>
> Not to be pedantic, but I suspect that s/he has conflated "philosophy" with
> "new age", as much of science owes itself to philosophy.
>
>
>
> marcos
>
>
>
> ============================================================
>
> FRIAM Applied Complexity Group listserv
>
> Meets Fridays 9a-11:30 at cafe at St. John's College
>
> lectures, archives, unsubscribe, maps at http://www.friam.org
>
>
>
>
>
>
>
> ============================================================
>
> FRIAM Applied Complexity Group listserv
>
> Meets Fridays 9a-11:30 at cafe at St. John's College
>
> lectures, archives, unsubscribe, maps at http://www.friam.org
>
>  ============================================================
> FRIAM Applied Complexity Group listserv
> Meets Fridays 9a-11:30 at cafe at St. John's College
> lectures, archives, unsubscribe, maps at http://www.friam.org
>
> Eric Charles
>
> Professional Student and
> Assistant Professor of Psychology
> Penn State University
> Altoona, PA 16601
>
>
>
> ============================================================
> FRIAM Applied Complexity Group listserv
> Meets Fridays 9a-11:30 at cafe at St. John's College
> lectures, archives, unsubscribe, maps at http://www.friam.org
>
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
lectures, archives, unsubscribe, maps at http://www.friam.org

Reply via email to