Personally, I like #3 best.
~PM

> Date: Mon, 3 Mar 2014 10:45:41 -0500
> Subject: Re: [agi] "Reward" and "utility" are fundamentally the same
> From: [email protected]
> To: [email protected]
> 
> There are different kinds of reinforcement learning.
> 
> 1. The AIXI model. The agent does not know the utility function and
> must learn it. It assumes the simplest model that fits observation.
> 
> 2. The MIRI model. A powerful agent lives in a complex environment
> with a simple and well understood (but poorly designed) utility
> function. It uses reasoning and thought experiments to predict which
> actions will maximize future reward.
> 
> 3. The animal model (including humans). A reward (or penalty) acts to
> increase (or decrease) the frequency of behavior performed at time t
> before the signal with effect proportional to 1/t.
> 
> 4. The practical AI model. The AI has no goals. Instead, its behavior
> is continually updated by the humans controlling it to meet the
> complex and poorly understood goals of the humans.
> 
> -- 
> -- Matt Mahoney, [email protected]
> 
> 
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
                                          


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to