On 15.04.2011 21:44 meekerdb said the following:
Entropy and information are related. In classical thermodynamics the
relation is between what constraint you impose on the substance and
dQ/T. You note that it is calculated assuming constant pressure -
that is a constraint; another is assuming
Hi Colin,
Energy cost is due to erasure of information only (Landauer
principle), and you can compute without erasing anything, as you need
to do if you do quantum computation. You might search on Landauer,
Bennett, Zurek, and on the Maxwell daemon.
Bruno
On 15 Apr 2011, at 02:27,
Colin,
I used to work in chemical thermodynamics for awhile and I give you the
answer from such a viewpoint. As this is the area that I know, then my
message will be a bit long and I guess it differs from the viewpoint of
people in information theory.
CLASSICAL THERMODYNAMICS
First entropy
Entropy and information are related. In classical thermodynamics the
relation is between what constraint you impose on the substance and
dQ/T. You note that it is calculated assuming constant pressure - that
is a constraint; another is assuming constant energy. In terms of the
phase space
Slizard did a whole bunch of stuff in this area in the 1940s. Feynmann
has some good introductions to it in his Lectures in Physics series (I
forget which volume), IIRC. This was more focussed on the
thermodynamics of computation (eg what efficiency limits are there on
processing bits).
Later on,
5 matches
Mail list logo