The Optimal Choice of Hypothesis Is the Weakest, Not the Shortest
<https://arxiv.org/abs/2301.12987>
Michael Timothy Bennett
<https://arxiv.org/search/cs?searchtype=author&query=Bennett,+M+T>

If A and B are sets such that A⊂B, generalisation may be understood as the
inference from A of a hypothesis sufficient to construct B. One might infer
any number of hypotheses from A, yet only some of those may generalise to B.
How can one know which are likely to generalise? One strategy is to choose
the shortest, equating the ability to compress information with the ability
to generalise (a proxy for intelligence). We examine this in the context of
a mathematical formalism of enactive cognition. We show that compression is
neither necessary nor sufficient to maximise performance (measured in terms
of the probability of a hypothesis generalising). We formulate a proxy
unrelated to length or simplicity, called weakness. We show that if tasks
are uniformly distributed, then there is no choice of proxy that performs
at least as well as weakness maximisation in all tasks while performing
strictly better in at least one. In experiments comparing maximum weakness
and minimum description length in the context of binary arithmetic, the
former generalised at between 1.1 and 5 times the rate of the latter. We
argue this demonstrates that weakness is a far better proxy, and explains
why Deepmind's Apperception Engine is able to generalise effectively.

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/T78fb8d90b9a51bf0-Me9085471b2cc10e29268313a
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to