[agi] Re: Motivational Systems of an AI

2006-12-03 Thread Pei Wang

Matt,

I have to argue against most of your conclusions (though some do make sense).

On 12/2/06, Matt Mahoney [EMAIL PROTECTED] wrote:


I argue this has to be the case because an intelligent system
cannot be allowed to modify its motivational system.


Disagree, if motivational system means the CONTENTS of goals,
motives, and drives.

See the following.


Our most fundamental models of intelligent agents require this (e.g. AIXI --
the reward signal is computed by the environment).


That to me is a major problem of AIXI. I would rather say that we
interpret some of our experience of the environment as reward signals
--- they are not objective, and our interpretation can be wrong.


 You cannot turn off hunger or pain.


To a degree we can --- by shifting attention to other things.


You cannot control your emotions.


I can, to a degree. ;-)


Since the synaptic weights cannot be altered by training (classical or operant
conditioning), they must be hardwired as determined by your DNA.


We don't know that as a fact. Even if it is, I don't think it means
that much for AI.


Do you agree?  If not, what part of this argument do you disagree with?


So I don't. Let's see.


That reward and punishment exist and result in learning in humans?


Of course reward and punishment exist, but not as a given. Instead,
these signals are produced by the system itself, based on experience.
The difference from your opinion is that certain experience may be
interpreted differently in different context, with respect to its
nature (reward or punishment), strength, and object (to which
operation or goal).


That there are neurons dedicated to computing reinforcement signals?


I don't know, and don't care too much.


That the human motivational system (by which I mean the logic of computing the
reinforcement signals from sensory input) is not trainable?


Of course the system must have innate mechanism for motivation, but we
need to be careful about what we mean by logic. In the case of NARS,
all initial goals are given to the system (via its interaction with
the environment), but all derived goals depend on the system's
beliefs, which comes from the system's own experience. The derivation
rules and functions are not trainable, but the contents and strength
of goals and motives are trainable. In general, the system often needs
to learn whether a signal is a reward.


That the motivational system is completely specified by DNA?


Again, it depends on what you mean by motivational system --- the
mechanism is innate, but the content is not.


That all human learning can be reduced to classical and operant conditioning?


In principle and at a very general level, I agree, but it won't help
us in designing AGI. As soon as the restriction of knowledge and
resources are taken into consideration, we cannot treat all learning
as conditioning. One concrete reason is that we cannot assume the
availability of reliable and immediate feedback on each
decision/behavior.


That humans are animals that differ only in the ability to learn language?


More than that. Even the ability to learn language may not be a
independent capability, but is based on other capabilities.


That models of goal seeking agents like AIXI are realistic models of 
intelligence?


Yes for goal seeking agents, but no for like AIXI. To me, AIXI is
an interesting idealized model of intelligence, but by no means
realistic.


Do you object to behavioralism because of their view that consciousness and
free will do not exist, except as beliefs?


We don't need to go that far in this discussion.


Do you object to the assertion that the brain is a computer with finite memory
and speed?  That your life consists of running a program?  Is this wrong, or
just uncomfortable?


That is right in principle, but not useful for the design of AGI,
similar to assertions like An intelligent system consists of atoms.

Pei

-
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303


[agi] Re: Motivational Systems of an AI

2006-12-03 Thread Richard Loosemore



Matt Mahoney wrote:

--- Richard Loosemore [EMAIL PROTECTED] wrote:
I am disputing the very idea that monkeys (or rats or pigeons or humans) 
have a part of the brain which generates the reward/punishment signal 
for operant conditioning.


This is behaviorism.  I find myself completely at a loss to know where 
to start, if I have to explain what is wrong with behaviorism.


Call it what you want.  I am arguing that there are parts of the brain (e.g.
the nucleus accumbens) responsible for reinforcement learning, and
furthermore, that the synapses along the input paths to these regions are not
trainable.  I argue this has to be the case because an intelligent system
cannot be allowed to modify its motivational system.  Our most fundamental
models of intelligent agents require this (e.g. AIXI -- the reward signal is
computed by the environment).  You cannot turn off hunger or pain.  You cannot
control your emotions.  Since the synaptic weights cannot be altered by
training (classical or operant conditioning), they must be hardwired as
determined by your DNA.



Pei has already spoken eloquently on many of these questions.


Do you agree?  If not, what part of this argument do you disagree with?

That reward and punishment exist and result in learning in humans?


Reward and punishment at what level?  Your use of behaviorist phrasing 
implies that you mean one particular interpretation of these terms, but 
there are others.  If it is the former, then the terms are so incoherent 
as mechanisms that there is no answer:  there simply is nothing as crude 
as behaviorist style reward and punishment going on.  As an idea bout 
mechanism it is bankrupt.


If you mean it at some other level, or if you mean the terms to be 
interpreted so generally that they could mean, for example, that there 
are mechanisms responsible for relaxation pressures that go in a 
particular direction, then of course they result in learning.




That there are neurons dedicated to computing reinforcement signals?


Similar answer to the previous.  Reinforcement signals could mean just 
about anything, but if you mean in the behaviorist sense, then there is 
no such thing as reinforcement learning going on.  And to understand 
*that* statement (the one I just made) you meed to understand a long 
story about why behaviorism is wrong.




That the human motivational system (by which I mean the logic of computing the
reinforcement signals from sensory input) is not trainable?


Now you are asking a question based on terms that (see above) are either 
ambiguous or incoherent.


If I back off from you interpretation of the motivational system, I can 
answer that the latter is probably a complicated entity, with many 
components, so some parts of it are trainable, and some others are not.




That the motivational system is completely specified by DNA?


This is a meaningless question.  Do you mean *directly* specified by the 
DNA?  Or do you mean that the DNA specifies a generator that builds the 
motivational system?  Or that the DNA specifies a generator that 
eventually builds the motivational system, after morphing through 
several intermediate mechanisms?  Are you allowing or excluding the 
interaction of the generators with the environment when they build the 
motivational system?  In all except the first case, the DNA only 
specifies things indirectly, so the phrase completely specified by DNA 
is ambivalent at best.





That all human learning can be reduced to classical and operant conditioning?


Of course I am disputing this.  This is the behaviorist idea that has 
been completely rejected by the cognitive science community since 1956.


If you are willing to bend the meaning of the terms classical and 
operant conditioning sufficiently far from their origins, you might be 
able to make the idea more plausible, but that kind of redefinition is a 
little silly, and I don't see you trying to do that.




That humans are animals that differ only in the ability to learn language?


Do I disagree with this?  Of course.  Humans are not moluscs that talk.


That models of goal seeking agents like AIXI are realistic models of
intelligence?


AIXI is not a model of a goal seeking agent, it is a mathematical 
abstraction of a goal seeking agent.  Of course it has no value as a 
realistic model of intelligence.



Do you object to behavioralism because of their view that consciousness and
free will do not exist, except as beliefs?


I assume you mean behaviorism.  My objection to behaviorism has 
nothing to do with any of their claims about free will or consciousness. 
 I happen to think that their opinions on such matters were generally 
incoherent (though not entirely so), but they were in good company on 
that one, so no matter.





Do you object to the assertion that the brain is a computer with finite memory
and speed?  That your life consists of running a program?  Is this wrong, or
just uncomfortable?


Well, I'm glad you ended on a 

[agi] Re: Motivational Systems of an AI

2006-12-03 Thread Richard Loosemore

J. Storrs Hall, PhD. wrote:

On Friday 01 December 2006 23:42, Richard Loosemore wrote:


It's a lot easier than you suppose.  The system would be built in two
parts:  the motivational system, which would not change substantially
during RSI, and the thinking part (for want of a better term), which
is where you do all the improvement.


For concreteness, I have called these the Utility Function and World Model in 
my writings on the subject...


Well  I am avoiding Utility Function precisely because it has a 
specific meaning in the context of the type of AI that I have been 
lambasting as the goal stack approach to motivation.


A plan that says Let RSI consist of growing the WM and not the UF suffers 
from the problem that the sophistication of the WM's understanding soon makes 
the UF look crude and stupid. Human babies want food, proximity to their 
mothers, and are frightened of strangers. That's good for babies but a person 
with greater understanding and capabilities is better off (and the rest of us 
are better off if the person has) a more sophisticated UF as well.


I don't want to take the bait on your baby-motivation analogy because I 
do not believe the difference between human baby and adult is the same 
as the difference between adult AI and even-smarter-adult-AI.  Some, 
including myself, are of the opinion that there is a threshold of 
sentience above which things settle down a lot, so the AI would never 
look back on its earlier motivational system and call it crude and stupid.


Also, implicit in your description of the UF and WM are some ideas that 
I have been explicitly avoiding in my discussion of diffuse 
motivational systems.  That would make some of your points not applicable.


No time to spell it out right now.  If you look back at the root of this 
thread you might see why, or you can wait until I get the thing written 
up properly.




It is not quite a contradiction, but certainly this would be impossible:
  deciding to make a modification that clearly was going to leave it
wanting something that, if it wanted that thing today, would contradict
its current priorities.  Do you see why?  The motivational mechanism IS
what the system wants, it is not what the system is considering wanting.


This is a good first cut at the problem, and is taken by e.g. Nick Bostrom in 
a widely cited paper at http://www.nickbostrom.com/ethics/ai.html


Well, Nick Bostrum is not the origin of this idea:  it is kind of obvious.



The system is not protecting current beliefs, it is believing its
current beliefs.  Becoming more capable of understanding the reality
it is immersed in?  You have implicitly put a motivational priority in
your system when you suggest that that is important to it ... does that
rank higher than its empathy with the human race?

You see where I am going:  there is nothing god-given about the desire
to understand reality in a better way.  That is just one more
candidate for a motivational priority.


Ah, but consider: knowing more about how the world works is often a valuable 
asset to the attempt to increase the utility of the world, *no matter* what 
else the utility function might specify. 


Whoa:  increase the utility of the world?  Again, your terms do not 
map onto a viewpoint of motivation that dumps the idea of a crude UF. 
In essence, you have restated the idea that I was attacking:  that 
increase the utility of the world is a motivation that trumps others. 
 It is not necessarily the case that this is the system's primary 
motivation.



Thus, a system's self-modification (or evolution in general) is unlikely to 
remove curiosity / thirst for knowledge / desire to improve one's WM as a 
high utility even as it changes other things. 


Yes and no.  I am going to have to get back to you on this.

Here is an idea to try to fit into that worldview.  After the 
Singularity, I would love to go into a closed domain in which I get to 
live in a replica of 17th century England, growing up there from 
childhood with my memories put on ice for the duration of a (then 
normal) lifetime, and with the goal of experiencing what it would have 
been like to be a Natural Philosopher discovering the wonder of science 
for the first time.  I want to discover things that are known in this 
era, after temporarily removing them from my mind.


So I would be what?  Contradicting my utility function by 
deliberately removing knowledge?  Seeking to do what?  Get the knowledge 
back a different way?  Am I seeking knowledge, or just seeking a new 
experience?


I claim the latter:  but that idea of seeking new experience just does 
not map onto the kind of silly :-) utility functions that AI people play 
games with today.  They cannot even represent the goal of having 
interesting subjective experiences, as far as I can see.



Richard Loosemore


There are several such properties of a utility function that are likely to be 
invariant under self-improvement or evolution. It is by 

Re: [agi] Re: Motivational Systems of an AI

2006-12-03 Thread Matt Mahoney

--- Richard Loosemore [EMAIL PROTECTED] wrote:

 
 
 Matt Mahoney wrote:
  --- Richard Loosemore [EMAIL PROTECTED] wrote:
  I am disputing the very idea that monkeys (or rats or pigeons or humans) 
  have a part of the brain which generates the reward/punishment signal 
  for operant conditioning.
 
  This is behaviorism.  I find myself completely at a loss to know where 
  to start, if I have to explain what is wrong with behaviorism.
  
  Call it what you want.  I am arguing that there are parts of the brain
 (e.g.
  the nucleus accumbens) responsible for reinforcement learning, and
  furthermore, that the synapses along the input paths to these regions are
 not
  trainable.  I argue this has to be the case because an intelligent system
  cannot be allowed to modify its motivational system.  Our most fundamental
  models of intelligent agents require this (e.g. AIXI -- the reward signal
 is
  computed by the environment).  You cannot turn off hunger or pain.  You
 cannot
  control your emotions.  Since the synaptic weights cannot be altered by
  training (classical or operant conditioning), they must be hardwired as
  determined by your DNA.
 
 
 Pei has already spoken eloquently on many of these questions.

Yes, and I agree with most of his comments.  I need to clarify that the part
of the motivational system that is not trainable is the one that computes top
level goals such as hunger, thirst, pain, the need for sleep, reproductive
drive, etc.  I think we can agree on this.  Regardless of training, everyone
will get hungry if they don't eat.  You can temporarily distract yourself from
hunger, but a healthy person can't change this top level goal.  If this were
not true, obesity would not be such a problem, and instead you would see a lot
of people starving themselves to death.

I think the confusion is over learned secondary goals, such as seeking money
to buy food, or education to get a better job.  So in that context, I agree
with most of your comments too.

  That all human learning can be reduced to classical and operant
 conditioning?
 
 Of course I am disputing this.  This is the behaviorist idea that has 
 been completely rejected by the cognitive science community since 1956.
 
 If you are willing to bend the meaning of the terms classical and 
 operant conditioning sufficiently far from their origins, you might be 
 able to make the idea more plausible, but that kind of redefinition is a 
 little silly, and I don't see you trying to do that.

How about if I call them supervised and unsupervised learning?

Of course this is not helpful.  What I am trying to do is understand how
learning works in humans so it can be modeled in AGI.  Classical conditioning
(e.g. Pavlov) has a simple model proposed by Hebb in 1949.  If neuron A fires
followed by B after time t, then the weight from A to B is increased in
proportion to AB/t (where A and B are activation levels).  The dependence on A
and B has been used in neural models long before synaptic weight changes were
observed in animal brains.  The factor 1/t (for t greater than a few hundred
milliseconds) is supported by animal experiments.

The model for reinforcement learning is not so clear.  We can imagine several
possibilities.

1. The weights of a neural network are randomly and temporarily varied.  After
a positive reinforcement, the changes become permanent.  If negative, the
changes are undone or made in the opposite direction.

2. The neuron activation level of B is varied by adding random noise, dB. 
After reinforcment r after time t, the weight change from A to B is
proportional to A(dB)r/t.

3. There is no noise.  Let dB be the rate of increase of B.  The weight change
is proportional to A(dB)r/t.

4. (as pointed out by Philip Goetz)
http://www.iro.umontreal.ca/~lisa/pointeurs/RivestNIPS2004.pdf
The weight change is proportional to AB(r-p), where p is the predicted
reinforcement (trained by classical conditioning) and r is the actual
reinforcement (tri-Hebbian model).

And many other possibilities.  We don't know what the brain uses.  It might be
a combination of these.  From animal experiments we know that the learning
rate is proportional to r/t, but not much else.  From computer simulations, we
know there is no best solution because it depends on the problem.  So I would
like to see an answer to this question.  How does it work in the brain?  How
should it be done in AGI?


-- Matt Mahoney, [EMAIL PROTECTED]

-
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303