What?
On Sat, Dec 21, 2013 at 2:25 PM, Piaget Modeler <[email protected]>wrote: > Whoa!! > > ~PM > > ------------------------------ > Date: Sat, 21 Dec 2013 13:43:19 -0600 > Subject: Re: [agi] A Random Thought... > From: [email protected] > To: [email protected] > > > Ok, which dimension are you attempting to scale up? Triviality corresponds > to minimal representational capacity, both in the environment and the agent > operating within it. Human beings are (currently) at the other end of that > scale, with enormous representational capacity for dealing with a highly > complex environment. The more complex (non-trivial) the environment, the > greater the representational capacity required of agents operating within > it in order to effectively make decisions. It is this dimension that I am > looking at. > > Learning algorithms are easy to understand, design, and implement. They > are just solutions to optimization problems. I do not think learning itself > is where the bottleneck lies. Instead I look at the representational > systems underlying those learning algorithms. The simplest learning > algorithms operate over tables of choices. They tabulate expected returns > or error levels for each choice, over many repetitions, and gradually > settle on the choice(s) with the maximum expected return or minimum > expected error level. Adding layers of sophistication, we begin to see > context matter more and more: Conditional choices and statefulness result > in much more interesting and coherent behavior. Generalizing over choices > and conditions and actions to those that are similar, we see an additional > gain in coherency, with algorithms that can deal with new situations > robustly based on previous experience with other situations. > > What is needed is to increase the expressivity of the underlying > representational schemes used by learning algorithms. Moving up to the > representational complexity level of ontologies, episodic memory, etc., the > representational scheme becomes ever more capable. In order to reason about > things, we need to represent those things effectively. Once we have a fully > capable representational scheme -- a programmatic framework for the > representation of Meaning, in all its forms, with all its inherent > ambiguities -- we can begin writing learning algorithms to extract meaning > from the environment, generate rules for predicting arbitrary unobserved > phenomena from arbitrary observed phenomena, recombining meanings to > produce new ones, choosing contextually appropriate and meaningful > behavior, etc. There is no understanding without meaning, and there is no > intelligence without understanding. > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/23050605-2da819ff> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
