Ben Goertzel wrote:
Actually, Ben, AIXI and AIXI-tl are both formal systems; there is no
internal component in that formal system corresponding to a "goal
definition", only an algorithm that humans use to determine when and how
hard they will press the reward button.
Well, the definitions of AIXI and AIXItl assume the existence of a "reward
function" or "goal function" (denoted V in the paper).

The assumption of the math is that this reward function is specified
up-front, before AIXI/AIXItl starts running.

If the reward function is allowed to change adaptively, based on the
behavior of the AIXI/AIXItl algorithm, then the theorems don't work anymore,
and you have a different sort of "synergetic" system such as Bill Hibbard
was describing.

If human feedback IS the reward function, then you have a case where the
reward function may well change adaptively based on the AI system's
behavior.

Whether the system will ever achieve any intelligence at all then depends on
how clever the humans are in doing the rewarding... as i said, Hutter's
theorems about intelligence don't apply...
Huh. We may not be on the same page. Using:
http://www.idsia.ch/~marcus/ai/aixigentle.pdf

Page 5:

"The general framework for AI might be viewed as the design and study of intelligent agents [RN95]. An agent is a cybernetic system with some internal state, which acts with output yk on some environment in cycle k, perceives some input xk from the environment and updates its internal state. Then the next cycle follows. We split the input xk into a regular part x0k and a reward rk, often called reinforcement feedback. From time to time the environment provides non-zero reward to the agent. The task of the agent is to maximize its utility, defined as the sum of future rewards."

I didn't see any reward function V defined for AIXI in any of the Hutter papers I read, nor is it at all clear how such a V could be defined, given that the internal representation of "reality" produced by Solomonoff induction is not fixed enough for any reward function to operate on it in the same way that, e.g., our emotions bind to our own standardized cognitive representations.

--
Eliezer S. Yudkowsky http://singinst.org/
Research Fellow, Singularity Institute for Artificial Intelligence

-------
To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]

Reply via email to