--- On Sun, 11/30/08, Philip Hunt <[EMAIL PROTECTED]> wrote:

> Can someone explain AIXI to me?

AIXI models an intelligent agent interacting with an environment as a pair of 
interacting Turing machines. At each step, the agent outputs a symbol to the 
environment, and the environment outputs a symbol and a numeric reward signal 
to the agent. The goal of the agent is to maximize the accumulated reward.

Hutter proved that the optimal solution is for the agent to guess, at each 
step, that the environment is simulated by the shortest program that is 
consistent with the interaction observed so far.

Hutter also proved that the optimal solution is not computable because the 
agent can't know which of its guesses are halting Turing machines. The best it 
can do is pick numbers L and T, try all 2^L programs up to length L for T steps 
each in order of increasing length, and guess the first one that is consistent. 
If there are no matches, then it needs to choose larger L and T and try again. 
That solution is called AIXI^TL. It's time complexity is O(T 2^L). In general, 
it may require L up to the length of the observed interaction (because there is 
a fast program that outputs the agent's observations from a list of length L).

In a separate paper ( http://www.vetta.org/documents/ui_benelearn.pdf ), Legg 
and Hutter propose defining universal intelligence as the expected reward of an 
AIXI agent in random environments.

The value of AIXI is not that it solves the general intelligence problem, but 
rather it explains why the problem is so hard. It also justifies a general 
principle that is already used in science and in practical machine learning 
algorithms: to choose the simplest hypothesis that fits the data. It formally 
defines "simple" as the length of the shortest program that outputs a 
description of the hypothesis.

For example, to avoid overfitting in neural networks, you should use the 
smallest number of connections and the least amount of training needed to fit 
the training data, then stop. In this case, the complexity of your neural 
network is the length of the shortest program that outputs the configuration of 
your network and its weights. Even if you don't know what that program is, and 
haven't chosen a programming language, you may reasonably expect that fewer 
connections, smaller weights, and coarser weight quantization will result in a 
shorter program.

-- Matt Mahoney, [EMAIL PROTECTED]




-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06
Powered by Listbox: http://www.listbox.com

Reply via email to