From: Fis [mailto:fis-boun...@listas.unizar.es] On Behalf Of Loet Leydesdorff
Sent: July 26, 2015 8:50 AM
To: 'Joseph Brenner'; 'Fernando Flores'; fis@listas.unizar.es
Subject: Re: [Fis] Answer to the comments made by Joseph
Dear Joe,
a) information is more than order; there is information
Dear John and colleagues,
So fundamentally we are talking about the same basic thing with information
and entropy.
The problem is fundamentally: the two are the same except for a constant.
Most authors attribute the dimensionality to this constant (kB).
From the perspective of
Of Loet
Leydesdorff
Sent: July 27, 2015 7:10 PM
To: John Collier; 'Joseph Brenner'; 'Fernando Flores'; fis@listas.unizar.es
Subject: RE: [Fis] Answer to the comments made by Joseph
Dear John and colleagues,
So fundamentally we are talking about the same basic thing with information and
entropy
Folks
I know there is a long legacy of equating information with entropy, and
dimensionally, they are the same. Qualitatively, however, they are
antithetical. From the point of view of statistical mechanics, information
is a *decrease* in entropy, i.e., they are negatives of each other.
This all
Dear Joe,
a) information is more than order; there is information in absence
(Deacon), in disorder, in incoherence as well as coherence;
The absent options provide the redundancy; that is, the complement of the
information to the maximal information [H(max)].
See also my recent
Dear Fernando,
This is becoming very interesting. I understand your critique, but I do not
believe it applies exactly to what I am trying to say. I start from a position
that the apodictic statement by Wiener is not or in any case is no longer
valid. In my view, the following should be taken