Folks

I know there is a long legacy of equating information with entropy, and
dimensionally, they are the same. Qualitatively, however, they are
antithetical. From the point of view of statistical mechanics, information
is a *decrease* in entropy, i.e., they are negatives of each other.

This all devolves back upon the requirement that *both* entropy and
information require a reference state. (The third law of thermodynamics.)
Once a reference distribution has been identified, one can then quantify
both entropy and information. It actually turns out that against any
reference state, entropy can be parsed into two components, mutual
information and conditional (or residual) entropy. Change the reference
state and the decomposition changes.
<http://people.clas.ufl.edu/ulan/files/FISPAP.pdf> (See also Chapter 5 in
<http://people.clas.ufl.edu/ulan/publications/ecosystems/gand/>.)

Cheers to all,
Bob

> Folks,
>
> Doing dimensional analysis entropy is heat difference divided by
> temperature. Heat is energy, and temperature is energy per degree of
> freedom. Dividing, we get units of inverse degrees of freedom. I submit
> that information has the same fundamental measure (this is a consequence
> of Scott Muller¡¯s asymmetry principle of information. So fundamentally we
> are talking about the same basic thing with information and entropy.
>
> I agree, though, that it is viewed from different perspectives and they
> have differing conventions for measurement.
>
> I agree with Loet¡¯s other points.
>
> John


_______________________________________________
Fis mailing list
Fis@listas.unizar.es
http://listas.unizar.es/cgi-bin/mailman/listinfo/fis

Reply via email to