Re: [Fis] Info Theory

2011-01-24 Thread Gavin Ritz


Hi there Stan
  
SS: Info theory presumably applies to everything and anything. 

GR: It was never intended to apply to anything but communication instruments. 
That is sending English language down a pipe.

GR: In my opinion it still only does, I cant get my head around how say 
information theory actually applies to direct human communication or organic 
sensing systems.  All our sensing systems are energy transduction systems, once 
inside the individual it 's moved via Na/K pumps aided by ADP to ATP 
conversions 
to the brain all electrical, chemical energy. So in the environment it's just a 
sound (phonon) or light (photon) or chemical or heat energy where are the bits 
(information theory part) or markers. They are just not there. 


Unless this information is what underlies energy and is what makes up the rest 
of the universe including dark matter and dark energy. And is also what 
underlies the theory of Geometricdynamics.(Relativity theory).?? How so I 
would not know.

 Gavin


 


On Sun, Jan 23, 2011 at 7:39 AM, Gavin Ritz garr...@xtra.co.nz wrote:

Are you saying Karl that Information theory is the glue that binds energy and 
entropy production?
or the the fabric behind these two concept?
If so what is the bridging qualitative and quantitative propositions and 
formulae for this binding?

It's quite something to say this, because one of the qualitative foundations 
of 
information theory is word frequency of English from Zipfs law. John Pierce 
(Information Theory)
Regards
Gavin







 From: joe.bren...@bluewin.ch  joe.bren...@bluewin.ch
To: karl.javors...@gmail.com; Pedro C. Marijuan pcmarijuan.i...@aragon.es; 
fis@listas.unizar.es
Sent: Sat, 22 January, 2011 7:32:24 AM
Subject: Re: [Fis] Info Theory


Dear Karl,   
 
The assumption I would like to check that we share is that existence  and 
energy 
are primitive and numbers something derived.  When one moves from the quantum 
vacuum or singularity into the thermodynamic world, as soon as change occurs, 
something is no longer totally itself; there is something new along side of it 
in 4D space-time. The number of entities has increased, and this is the 
situation is the reality of which addition is the model. Iteration, which also 
occurs in reality, does the rest. If I understand you correctly, you feel that 
numbers, once available and manipulated in more complex ways, can model many 
other things, especially, of course, aspects of information.
 
If a numerical perspective is convenient and even necessary for an 
understanding 
of nature, I would still like to know if it is sufficient. Are you able to 
capture, in your information theory, for example, the informational processes 
involved in:
 
· emotions
· creativity
· anti-social behavior (rational and irrational)
· complex political processes
· your own theory?
 
I think it would make for a more interesting and productive discussion if you 
were to tell us where your theory does NOT apply, rather than let us raise 
naïve 
objections to which you already have clear answers. I would like to know, for 
example, which of several possible approaches to the definition of a logical 
object are involved; at what point the limitations of machines become 
determining; and under what conditions one should seek to maximize (because 
valuable) heterogeneity as  opposed to homogeneity. Very interesting 
discussions 
can then be envisaged at the “boundaries” between different approaches.
 
Thank you and best wishes,
 
Joseph

Ursprüngliche Nachricht
Von: karl.javors...@gmail.com
Datum: 20.01.2011 21:03
An: Jerry Chandlerjerry_lr_chand...@mac.com, Joseph 
Brennerjoe.bren...@bluewin.ch, Pedro C. 
Marijuanpcmarijuan.i...@aragon.es
Betreff: Info Theory

Hope that the FIS server will eventuially accept this, too.. For you, 
individually:


Information Theory:
Let me answer the points raised so far:
Joe Brenner:
My hope is that this discussion will have a good deal to do with qualitative 
as 
well as quantitative aspects of information. Perhaps people should state 
clearly 
what the primary interests and objectives are of their remarks. 





Jerry Chandler:
The unspoken premise of many discussants appears to me to be a view of 
information theory as a universal glue, a universal predicate, a universal 
code.
The assertion is outspoken, explicit and apodictically declaratory: 
information 
theory IS a universal glue, a universal predicate, a universal code

Yet, any effort to use quantum logic to describe inheritance requires the 
construction of semantic bridges between  messages before the encoding 
occurs. 
The existence of such semantic links or connections is intrinsic to the 
logical 
premise or assertion lies in the encoding process, not the experimental 
science 
that generates the information.
The concepts and procedures underlying quantum logic and inheritance root 
BOTH 
in a common concept of 

Re: [Fis] Info Theory

2011-01-24 Thread Gavin Ritz
Karl
I cant fault your thinking, ( with some very important things you mention) but 
your comment about Shannon's view doesn't make a whole lot of sense to me, 
fire- 
or not fire is just one logical connective form. Just One aspect of proposition 
logic connective. Imperative Logic which is actually more pervasive in organic 
systems is still unaccounted by your comments below.

In fact Imperative Logic (fire neuron now) is how organic systems communicate 
with itself and others. And totally unaccounted for in most models.

Shannon's communication theory is about getting data down a pipe and reading it 
at the other end. Not really intended to be anything else.

If you say Limits of Glue then Information Theory does not underlie entropy 
production (or energy) or the Non-equilibrium free energy (combination of the 
two), and in fact may only be one qualitative aspect of entropy production. 
Then 
cannot be used to explain the Reality of Nature.

Regards
Gavin






From: karljavorszky karl.javors...@gmail.com
To: Gavin Ritz garr...@xtra.co.nz
Cc: joe.bren...@bluewin.ch; fis@listas.unizar.es; pcmarijuan.i...@aragon.es
Sent: Mon, 24 January, 2011 1:58:28 AM
Subject: Re: [Fis] Info Theory

Limits of Glue

Joe:...that existence and energy are primitive and numbers something derived.

Yes of course. We know that Nature exists and has manifold properties.
(Thomas Aquinas).
We speak about our experiences with Nature. To make certain that we
understand each other clearly, we use words with progressive degrees
of formal meaning. The extreme of this is that we use the public
language, i.e. numbers, - where no person has (should have) subjective
connotations, and the denotations of the words are clear. The imagery
built up by this method has the shortcomings that it is a very
abstract, detached, idealised way of speaking about Nature. It has the
advantages that we each know that we mean the same as we say in this
model Nature is in a constant change as we refer to the fact that the
Euclid spaces which give mass a localisation are derived from the
concept that a reordering always takes place, no side of a logical
argument having any innate, intrinsic claim of being more true than
other aspects. It is a continuous reordering which brings forth the
convoys of objects moving together (“strings”) and one of the readings
yields coordinates in two perfectly rectangular spaces. So the basic
principle is that it moves, as Heraclit said it should.

What I say in normal, subjectively colored language is that space is
actually two spaces which are merged into each other. The fabric of
space is made up of the undecided logical (sub-)questions of the
relevance of aspects. If it is more descriptive of a+b=c that 2a-3b is
in such and such way more related to b-2a than to 2b-3a (just to
mention an example), then space either constricts or expands or the
strings going thru the truth points of this debate have to carry more
fillings or less. The stuff must be somewhere. The 4D space you ask
about is perfectly there, with strings attached, twice.

Yes, physiology is a science of accounting and maintaining very strict
limits. This is even more true of neurology. That we humans have funny
ideas is built into the mechanism and can be seen e.g. on wolves,
bears, apes as they play and chase imaginary prey (which is strictly
speaking a hallucination).

The translation sequenced-commutative is what we see in the DNA and in
the functions of the brain. The electrical discharges which we call
thoughts are sequenced and come from specific places, but are
otherwise uniform. The cells fire or fire not. They have two logical
states. This is the Shannon way of doing things. Then, interdependent
with this, we have multiform material which is displaced. The fluids
are only generally somewhere in the region, they can lose their place,
and quite importantly they are of several varieties. The anti-Shannon
idea is that there are more forms in Nature (which we can speak about
in a formalized fashion) than this one and not this one.

The model presented is not an explanation for everything and all. It
is a tool to play with. We have 16 kinds of building blocks in two
sets, black and white. We pair the blocks and order them. Then we
reorder them again. We then discuss which pair goes with which other
pairs together in a convoy. This appears at first sight very
complicated but is extremely logical.

The glue in question connecting and partly fusing concepts in our
brain and between sciences and societies and among particles and
galaxies is well pictured in the formal language by the strings that
show the (possibly irrelevant) spatial coordinates of the convoys. It
is not the accountant’s job to give names to amounts systematically
under way and partly misplaced. It is the scientists’ prerogative to
decide what they call a string, a field, a force, a molecule.
Accounting processes connect points in Euclid spaces with extents. We

Re: [Fis] Info Theory

2011-01-24 Thread Gavin Ritz


Hi there Loet
If the information exchange is provided with meaning, then this is probably 
just 
a qualitative aspect of entropy production. However this is not Information 
Theory.

The Relation between energy and information theory is 0.693kT (k is the 
Boltzmann constant and T is the temp in kelvin) Joules/bit this however is for 
a 
machine not a living organism.

Knowledge transfer systems are Imperative Logic Systems hereto totally 
uncounted 
for.

Regards
Gavin



Dear colleagues, 
 
It seems to me that the relation between information and energy provides the 
special case that the entropy is thermodynamic entropy. The relation is S = 
k(B) 
H. H is dimensionless, but S is not because k(B) adds the dimensionality of 
Joule/Kelvin. H can also be considered as probabilistic entropy. S is relevant 
in the case that the system of reference is the chemico-physical one based on 
collisions among particles. This level – the exchange of momenta and energy – 
is 
always involved in higher-order exchange processes, but the next-order ones 
emerge on top of the lower-order ones. 

 
For example, when specifically molecules are exchanged, life can emerge 
(Maturana). The self-organization may also reduce the uncertainty locally 
(“negentropy”). The system of reference, however, then is different from the 
chemico-physical one. The information exchange is provided with meaning.
 
Things change dramatically when meaning can again be communicated because then 
models can be entertained at a more rapid speed than the underlying (that is, 
modeled) systems. The redundancy generation can then prevail over the entropy 
generation and a knowledge-based economy, for example, maintained. The 
discursive models proliferate options other than the ones which occurred 
historically. This cultural system incurs on the historical manifestations and 
thus counteracts upon their following of the entropy law. The social system, 
for 
example, can be based on other premises than the lower-order ones. For example, 
the “survival of the fittest” can be replaced by universal human rights.
 
In other words: the specification of the system of reference provides the 
information exchanges with meaning. This meaning can again be communicated 
reflexively in the respective disciplines. The systems can be expected to gain 
in their capacity to process complexity insofar as these different layers 
become 
more nearly decomposable. This expansion spans the different dimensionalities 
and thus can be expected to enlarge the space for knowledge-based 
interventions. 

 
Best wishes, 
Loet
 



LoetLeydesdorff
Professor, University of Amsterdam
AmsterdamSchoolof Communications Research (ASCoR), 
Kloveniersburgwal 48, 1012 CXAmsterdam. 
Tel.: +31-20- 525 6598; fax: +31-842239111
l...@leydesdorff.net ; http://www.leydesdorff.net/
 
From:fis-boun...@listas.unizar.es [mailto:fis-boun...@listas.unizar.es] On 
Behalf Of Gavin Ritz
Sent: Monday, January 24, 2011 12:42 PM
To: Stanley N Salthe
Cc: fis@listas.unizar.es
Subject: Re: [Fis] Info Theory
 
 
Hi there Stan
  
SS: Info theory presumably applies to everything and anything. 

GR: It was never intended to apply to anything but communication instruments. 
That is sending English language down a pipe.

GR: In my opinion it still only does, I cant get my head around how say 
information theory actually applies to direct human communication or organic 
sensing systems.  All our sensing systems are energy transduction systems, once 
inside the individual it 's moved via Na/K pumps aided by ADP to ATP 
conversions 
to the brain all electrical, chemical energy. So in the environment it's just a 
sound (phonon) or light (photon) or chemical or heat energy where are the bits 
(information theory part) or markers. They are just not there. 


Unless this information is what underlies energy and is what makes up the rest 
of the universe including dark matter and dark energy. And is also what 
underlies the theory of Geometricdynamics.(Relativity theory)..?? How so I 
would not know.

 Gavin
 
 
 
On Sun, Jan 23, 2011 at 7:39 AM, Gavin Ritz garr...@xtra.co.nz wrote:
Are you saying Karl that Information theory is the glue that binds energy and 
entropy production?
or the the fabric behind these two concept?
If so what is the bridging qualitative and quantitative propositions and 
formulae for this binding?

It's quite something to say this, because one of the qualitative foundations of 
information theory is word frequency of English from Zipfs law. John Pierce 
(Information Theory)
Regards
Gavin
 
 



From:joe.bren...@bluewin.ch joe.bren...@bluewin.ch
To: karl.javors...@gmail.com; Pedro C. Marijuan pcmarijuan.i...@aragon.es; 
fis@listas.unizar.es
Sent: Sat, 22 January, 2011 7:32:24 AM
Subject: Re: [Fis] Info Theory
Dear Karl,   
 
The assumption I would like to check that we share is that existence and energy 
are primitive and 

Re: [Fis] Info Theory

2011-01-24 Thread John Collier


At 02:05 PM 1/24/2011, Loet Leydesdorff wrote:
Content-Type:
multipart/alternative;

boundary==_NextPart_000_00C9_01CBBBC7.684CC9E0
Content-Language: en-us
Dear colleagues, 

It seems to me that the relation between information and energy provides
the special case that the entropy is thermodynamic entropy. The relation
is S = k(B) H. H is dimensionless, but S is not because k(B) adds the
dimensionality of Joule/Kelvin. H can also be considered as probabilistic
entropy. S is relevant in the case that the system of reference is the
chemico-physical one based on collisions among particles. This level –
the exchange of momenta and energy – is always involved in higher-order
exchange processes, but the next-order ones emerge on top of the
lower-order ones. 
I think that joules are energy, and temperature is energy per degrees of
freedom. If you cancel out the energy part you get degrees of freedom,
which is dimensionless (a number), and is also a good measure for things
like information in physical terms. So I don't see a problem. W

For example, when specifically molecules are exchanged, life can emerge
(Maturana). The self-organization may also reduce the uncertainty locally
(“negentropy”). The system of reference, however, then is different
from the chemico-physical one. The information exchange is provided with
meaning.
Only because it is interpreted (biosemiotics).

Things change dramatically when meaning can again be communicated because
then models can be entertained at a more rapid speed than the underlying
(that is, modeled) systems. The redundancy generation can then prevail
over the entropy generation and a knowledge-based economy, for example,
maintained. The discursive models proliferate options other than the ones
which occurred historically. This cultural system incurs on the
historical manifestations and thus counteracts upon their following of
the entropy law. The social system, for example, can be based on other
premises than the lower-order ones. For example, the “survival of the
fittest” can be replaced by universal human rights.

In other words: the specification of the system of reference provides the
information exchanges with meaning. 
Quite.
This meaning can again be
communicated reflexively in the respective disciplines. The systems can
be expected to gain in their capacity to process complexity insofar as
these different layers become more nearly decomposable. This expansion
spans the different dimensionalities and thus can be expected to enlarge
the space for knowledge-based interventions. 
Dimensionalities add degrees of freedom, and thus information capacity.
So information capacity can emerge (or even be created) by the sort of
process you mention. It is all physically grounded, though.
One more thing:
GR: That's thermodynamically impossible. Any organic system requires to
convert and transduce energy so you may think it does no work but the
relationship is like Ostwald's Ripening there is always an energy
cost
Sorry Gavin, but you are mistaken. The entropy budget is made at the
expense of information loss in these cases.
Incidentally, you are posting too often. The rules say two a week. This
allows people to check sources, etc. Google is good, plus archives of the
fis list.
Schroedinger, What is Life? (1945). The connection is via
negentropy, and then to biological information in the DNA (he called it a
nonperiodic crystal).
The 1929 Szillard, SZILARD L., Z. Phys., 53 (1929)
840-856, paper is in German. An English translation can be found in Leff
and Rex, Maxwell's Demon (1990, Princeton University Press). It is
generally regarded as the first explicit connection between information
and physics.
John





Professor John Collier, Acting HoS and Acting Deputy HoS

colli...@ukzn.ac.za
Philosophy and Ethics, University of KwaZulu-Natal, Durban 4041 South
Africa
T: +27 (31) 260 3248 / 260 2292 F:
+27 (31) 260 3031

http://collier.ukzn.ac.za/



___
fis mailing list
fis@listas.unizar.es
https://webmail.unizar.es/cgi-bin/mailman/listinfo/fis


Re: [Fis] Info Theory

2011-01-24 Thread Robert Ulanowicz
Dear Gavin,

With all due respect, I would like to strongly disagree. I believe the  
communications format is an historical artifact that were best put  
aside. It unnecessarily constrains the application of information  
theory for many useful purposes.

Most especially, the calculus that is built upon the Shannon formula  
has incredibly wide application. It is useful anywhere constraint  
enters the picture.

Of course, as a thermodynamicist myself, I understand your desire for  
purity of definition. So if you were to insist on constraining  
information to communication, then something like Collier's  
enformation needs to be coined to handle the myriad of other  
productive ways that information theory can be of use to us.

I'll warn you, however, it's going to be difficult to draw a clear  
line between information and enformation. :)

The best,
Bob

-
Robert E. Ulanowicz|  Tel: +1-352-378-7355
Arthur R. Marshall Laboratory  |  FAX: +1-352-392-3704
Department of Biology  |  Emeritus, Chesapeake Biol. Lab
Bartram Hall 110   |  University of Maryland
University of Florida  |  Email u...@cbl.umces.edu
Gainesville, FL 32611-8525 USA |  Web http://www.cbl.umces.edu/~ulan
--


Quoting Gavin Ritz garr...@xtra.co.nz:



 Hi there Stan

 SS: Info theory presumably applies to everything and anything.

 GR: It was never intended to apply to anything but communication instruments.
 That is sending English language down a pipe.

 GR: In my opinion it still only does, I cant get my head around how say
 information theory actually applies to direct human communication or organic
 sensing systems.  All our sensing systems are energy transduction  
 systems, once
 inside the individual it 's moved via Na/K pumps aided by ADP to ATP  
 conversions
 to the brain all electrical, chemical energy. So in the environment  
 it's just a
 sound (phonon) or light (photon) or chemical or heat energy where  
 are the bits
 (information theory part) or markers. They are just not there.


 Unless this information is what underlies energy and is what makes  
 up the rest
 of the universe including dark matter and dark energy. And is also what
 underlies the theory of Geometricdynamics.(Relativity theory).?? How so I
 would not know.

  Gavin





 On Sun, Jan 23, 2011 at 7:39 AM, Gavin Ritz garr...@xtra.co.nz wrote:

 Are you saying Karl that Information theory is the glue that binds energy and
 entropy production?
 or the the fabric behind these two concept?
 If so what is the bridging qualitative and quantitative propositions and
 formulae for this binding?

 It's quite something to say this, because one of the qualitative  
 foundations of
 information theory is word frequency of English from Zipfs law. John Pierce
 (Information Theory)
 Regards
 Gavin






 
  From: joe.bren...@bluewin.ch  joe.bren...@bluewin.ch
 To: karl.javors...@gmail.com; Pedro C. Marijuan pcmarijuan.i...@aragon.es;
 fis@listas.unizar.es
 Sent: Sat, 22 January, 2011 7:32:24 AM
 Subject: Re: [Fis] Info Theory


 Dear Karl,

 The assumption I would like to check that we share is that  
 existence  and energy
 are primitive and numbers something derived.  When one moves from  
 the quantum
 vacuum or singularity into the thermodynamic world, as soon as  
 change occurs,
 something is no longer totally itself; there is something new along  
 side of it
 in 4D space-time. The number of entities has increased, and this is the
 situation is the reality of which addition is the model. Iteration,  
 which also
 occurs in reality, does the rest. If I understand you correctly,  
 you feel that
 numbers, once available and manipulated in more complex ways, can model many
 other things, especially, of course, aspects of information.

 If a numerical perspective is convenient and even necessary for an  
 understanding
 of nature, I would still like to know if it is sufficient. Are you able to
 capture, in your information theory, for example, the informational  
 processes
 involved in:

 · emotions
 · creativity
 · anti-social behavior (rational and irrational)
 · complex political processes
 · your own theory?

 I think it would make for a more interesting and productive  
 discussion if you
 were to tell us where your theory does NOT apply, rather than let  
 us raise naïve
 objections to which you already have clear answers. I would like to  
 know, for
 example, which of several possible approaches to the definition of  
 a logical
 object are involved; at what point the limitations of machines become
 determining; and under what conditions one should seek to maximize (because
 valuable) heterogeneity as  opposed to homogeneity. Very  
 interesting discussions
 can then be envisaged 

[Fis] Foundational Views of Shannon Information Theory

2011-01-24 Thread Jerry LR Chandler

List:

My responses to recent posts by Karl, Stan, Joe, Loet, Gavin, John, and Bob by 
the number of the digest that I rec’d. I seek to address several basic issues. 

 

First, I would repeat my assertion from my post of Jan. 20, 2010, along with 
Karl’s denial and my comments about his denials:

 

JLRC: The unspoken premise of many discussants appears to me to be a view of 
information theory as a  universal glue, a universal predicate, a universal 
code.

KJ: The assertion is outspoken, explicit and apodictically declaratory: 
information theory IS a universal glue, a universal predicate, a universal code

 

Karl: Out-spoken?

JLRC:  Yes, I spoke-out.  :-)

Karl: Explicit? 

JLRC: Yes. Rosen argues that biology requires a separate symbol system, that is 
outside of mathematical category theory. My explicit response to the category 
theory approach to information theory is contained in three recent papers – 
Axiomathes, Discrete Applied Math, and a chapter in a book by Vrobel and Otto 
Rossler. If desired, I will forward copies of these papers to list members.

Karl: apodictically declaratory: 

JLRC: Yes!  By design.   ;-) 

JLRC:  Perhaps you have not considered the reasons why Shannon information 
lacks universality. So, Precisely what is it that you are denying about the 
appearances of information theory?  

·  That category theory is applicable to biology?

·  By inference, that set theory / predicate logic is sufficient to 
describe optical isomers?

·  That the simple “yes/No” choice essential to Shannon information is a 
universal code for human knowledge?  (The notion of a binary encoding of all 
information is denied by Dalton’s premise – the ostensive source of chemical 
codes.)

·  Or, is it that you believe that addition is a universal operation of 
mathematics?

JLRC: Your numerous posts on your decade-long mediations on the nature of 
arithmetic remain unpersuasive. The consistency of group theory and ring theory 
provide an adequate explanation for all iterative arithmetic operations. Your 
persistence is admirable, your intelligence is substantial, your logic 
questionable and your conclusions lack extension. 

 

Stan (545:10) Re: [Fis] Ostension and the Chemical / Molecular Biological 
Science,   …It is this translation from material observations into logical 
form, in particular into fully explicit, crisp logical form that I am 
questioning.  Yes, it can lead to short term triumphs, via engineering,…

 

JLRC: Hu, I think you miss the point. The abstract symbol systems of 
Dalton, Lavoisier, and Coulomb underly the foundations of thermodynamics as 
well as the Shannon theory of information as well as our concept of such 
abstractions as “energy” and “entropy.” These symbol systems are now firmly 
embedded in the logic of scientific communications. Perhaps you wish to infer 
that concept of ostension is not useful in the natural sciences?  Or, is it 
that in your world view, “utility” is a bad word? 

BTW, Lavoisier / Daltonian logical forms are not fully explicit in the usual 
sense of mathematics. They are closer to codes with an exact syntax.

Joe (245:11) …”that existence and energy are primitive and numbers something 
derived.”

JLRC: Are you putting the cart before the horse? As a consequence of the 
international system of units, number takes priority over all other scientific 
and economic proper names. Number is the antecedent to expressing quantity of 
most any sort.

JLRC: Are you attempting to substitute semantics for syntax in your view of 
information theory? In your view of symbolic logic?  In your view of the 
concept of order?

JLRC: The order of the atomic numbers of the chemical elements stands in 
one:one correspondence with the any list of objects, with the listing of 
elements of a group, in a listing of the roots of a polynomial, in a listing of 
a vector, in a listing of the nodes of a graphs, and so forth. The existence of 
a listing is essential to the basic attributes of a message. It is essential to 
communication.

 

Joe (245:11) …and under what conditions one should seek to maximize (because 
valuable) heterogeneity as opposed to homogeneity.

 

JLRC: What fundamental classes of informational variables can be used to 
express heterogeneity?  Homogeneity? How do such classes relate to Rosen’s 
postulates of separate and distinct symbol systems? Or, Aristotelian causal 
structures?  Or, Descarte’s “clear and distinct” ideas

 

Gavin (245:12) …one of the qualitative foundations of information theory is 
word frequency of English from Zipfs law.

 

JLRC: Minor technical point.  Perhaps you mean the frequency of usage of 
different alphabet symbols in a linguistic message?

 

Stan (245:12) … Put otherwise, does anyone know of data about natural things 
that would not deliver a power law?

 

JLRC: Power laws are the exception, not the rule in the natural sciences. For 
example, catalysis, the source of nearly all of biological