Harshad RJ wrote:
On Feb 3, 2008 10:22 PM, Richard Loosemore <[EMAIL PROTECTED]
<mailto:[EMAIL PROTECTED]>> wrote:
Harshad RJ wrote:
> I read the conversation from the start and believe that Matt's
> argument is correct.
Did you mean to send this only to me? It looks as though you mean it
for the list. I will send this reply back to you personally, but let me
know if you prefer it to be copied to the AGI list.
Richard, thanks for replying. I did want to send it to the list.. and
your email address (as it turns out) was listed on the forum for
replying to the list.
> There is a difference between intelligence and motive which Richard
> seems to be ignoring. A brilliant instance of intelligence could
still
> be subservient to a malicious or ignorant motive, and I think that is
> the crux of Matt's argument.
With respect, I was not at all ignoring this point: this is a
misunderstanding that occurs very frequently, and I thought that I
covered it on this occasion (my apologies if I forgot to do so..... I
have had to combat this point on so many previous occasions that I may
have overlooked yet another repeat).
The crucial words are "... could still be subservient to a malicious or
ignorant motive."
The implication behind these words is that, somehow, the "motive" of
this intelligence could arise after the intelligence, as a completely
independent thing over which we had no control. We are so used to this
pattern in the human case (we can make babies, but we cannot stop the
babies from growing up to be dictators, if that is the way they happen
to go).
This implication is just plain wrong.
I don't believe so, though your next statement..
If you build an artificial
intelligence, you MUST choose how it is motivated before you can even
switch it on.
... might be true. Yes, a motivation of some form could be coded into
the system, but the paucity of expression in the level at which it is
coded, may still allow for "unintended" motivations to emerge out.
Say, for example, the motivation is coded in a form similar to current
biological systems. The AGI system is motivated to keep itself happy,
and it is happy when it has sufficient electrical energy at its disposal
AND when the pheromones from nearby humans are all screaming "positive".
It is easy to see how this kind of motivation could cause unintended
results. The AGI system could do dramatic things like taking over a
nuclear power station and manufacturing its own pheromone supply from a
chemical plant. Or it could do more subtle things like, manipulating
government policies to ensure that the above happens!
Even allowing for a higher level of coding for motivation, like those
Asimov's Robot rules (#1 : Though shall not harm any human), it is very
easy for the system to go out of hand, since such codings are ambiguous.
Should "stem cell research" be allowed for example? It might harm some
embryos but help more number of adults. "Should prostitution be
legalised?" It might harm the human gene pool in some vague way, or
might even harm some specific individuals, but it also allows the
victims themselves to earn some money and survive longer.
So, yes, motivation might be coded, but an AGI system would eventually
need to have the *capability* to deduce its own motivation, and that
emergent motivation could be malicious/ignorant.
I quote the rest of the message, only for the benefit of the list.
Otherwise, my case rests here.
Stepping back for a moment, I think the problem that tends to occur in
discussions of AGI motivation is that the technical aspects get
overlooked when we go looking for nightmare scenarios. What this means,
for me, is that when I reply to a suggestion such as the one you give
above, my response is not "That kind of AGI, and AGI behavioral problem,
is completely unimaginable", but instead what I have to say is "That
kind of AGI would not actually BE an AGI at all, because, for technical
reasons, you would never be able to get such a thing to be intelligent
in the first place".
There is a subtle difference between these two, but what I find is that
most people mistakenly believe that I am making the first kind of
reponse instead of the second.
So, to deal with your suggestion in detail.
When I say that some kind of motivation MUST be built into the system, I
am pretty much uttering a truism: an AGI without any kind of
motivational system is like a swimmer with no muscles. It has to be
driven to do something, so no drives mean no activity.
Putting that to one side, then, what you propose is an AGI with an
extremely simple motivational system: seek electricity and high human
pheromonal output.
I don't suggest that this is unimaginable (it is!), but what I suggest
is that you implicitly assume a lot of stuff that, almost certainly,
will never happen.
You assume that the system does not go through a learning phase
(childhood) during which it acquires its knowledge by itself. Why do
you assume this? Because an AGI that was motivated only to seek
electricity and pheromones is going to be as curious, as active, as
knowledge seeking, as exploratory (etc etc etc) as a moth that has been
preprogrammed to go towards bright lights. It will never learn aything
by itself because you left out the [curiosity] motivation (and a lot
else besides!). All you can hope for is that human programmers insert
all of its knowledge by hand .... and I mean all of its knowledge,
because it will have no motivation to acquire more (by your assumption
that it only seeks two very narrow goals).
Most AI people agree that a full AGI cannot be built if all its
knowledge has to be compiled by hand. It must be able to learn by itself.
But when we try to get an AGI to have the kind of structured behavior
necessary to learn by itself, we discover ..... what? That you cannot
have that kind of structured exploratory behavior without also having an
extremely sophisticated motivation system.
In other words you cannot have your cake and eat it too: you cannot
assume that this hypothetical AGI is (a) completely able to build its
own understanding of the world, right up to the human level and beyond,
while also being (b) driven by an extremely dumb motivation system that
makes the AGI seek only a couple of simple goals.
If an AGI could ever be built with only such simplistic goals then,
sure, that kind of motivational system would be ridiculously unstable,
and liable to blow up in your face. But you simply cannot do that: try
it and you will see. The thing may have isolated bursts of rationality
within some limited domain, if you go to the trouble hand coding it in
that domain, but if you try to persuade such a simply-motivated system
to actually grow and develop into an adult that would be your equal, you
will spend your whole life in absolute frustration as it just sits there
and looks up at you with all the intellectual depth of a Sphex Wasp or a
stick insect.
I believe that if you go and analyze every single science-fiction robot
nightmare scenario that was ever conceived, you will see the same pair
of unworkable assumptions The hypthetical AGI is supposed to be so
smart that it can understand all the incredible depths of complexity
that are involved in building a motivational system (because, by
assumption, it knows enough about its own design to be able to make
coherent modifications), and yet at the same time it is supposed to be
so simple underneath that it cannot understand anything except for a
primal desire to get electricity and a pheromone hit!
So, these science fiction AGIs would be able to take part in this
discussion that you and I are having right now, and go into such
incredible technical detail about the mechanics of building a
motivational system that neither of us could follow ....... but this
same AGI is also supposed to be following a simple explicit-rule-based
motivational system in which rule #1 is "Though shall not harm any
human", where this statement has to be interpreted?!
Now, the exact details of the line of argument I propose here are not
set in stone (there is much work to be done to see if the argument
really does play out, the way it seems to), but right now this
particular approach to AGI motivation is the only one that does not have
fundamental contradictions built into it.
There is more to say, but not in an already long post.
Richard Loosemore
Nature does this in our case (and nature is very
insistent that it wants its creations to have plenty of selfishness and
aggressiveness built into them, because selfish and aggressive species
survive), but nature does it so quietly that we sometimes think that all
she does is build an intelligence, then leave the motivations to grow
however they will. But what nature does quietly, we have to do
explicitly.
My argument was (at the beginning of the debate with Matt, I believe)
that, for a variety of reasons, the first AGI will be built with
peaceful motivations. Seems hard to believe, but for various technical
reasons I think we can make a very powerful case that this is exactly
what will happen. After that, every other AGI will be the same way
(again, there is an argument behind that). Furthermore, there will not
be any "evolutionary" pressures going on, so we will not find that (say)
the first few million AGIs are built with perfect motivations, and then
some rogue ones start to develop.
So, when you say that "A brilliant instance of intelligence could still
be subservient to a malicious or ignorant motive" you are saying
something equivalent to "Toyota could build a car with a big red button
on the roof, and whenever anyone slapped the button a nuclear weapon
would go off in the car's trunk." Technically, yes, I am sure Toyota
could find a way to do this! But oing this kind of thing is not an
automatic consequence (or even a remotely probably consequence) of a
company becoming a car manufacturer with enough resources to do such a
thing. Similarly, having malevolent motives is not an automatic
consequence (or even a remotely probably consequence) of a system being
an intelligence with enough resources to do such a thing.
> There are two possibilities:
> 1. The AGI in question could have been programmed to choose it's own
> motive, in which case, the AGI may very well choose a motive that is
> malicious to humanity.
As I just explained, it must already have a motive before it can do any
choosing. That first motive will determine whether it even considers
the possibility of being malevolent.
Also, to be smart enough to redesign its own motivations, it has to be
very smart indeed, having a profound understanding of different motives
and the consequences of tampering with its own motivations. Under those
circumstances I believe it will simply take steps to reinforce its
bening motives to make sure there is no chance of accidental deviation.
>
> 2. The AGI could be programmed to satisfy a motive specified by the
> creator, in which case the maliciousness (or ignorance) of the
creator
> is what should be considered in this discussion. And since the
> creators (that is we, humans) are known to be both ignorant and
> capable of malice, the system is highly susceptible to "an AGI
> singularity that leads to humanity being rendered redundant".
Again, this argument is similar to the "Since Toyota is staffed by
humans, who are are known to be both ignorant and capable of malice, the
Toyota company is highly susceptible to a scenario in which they create
a car with a nuclear weapon in the trunk". Being capable of doing it is
not the same as actually doing it.
All that needs to happen is for the first AGI to be built with benign
motives, and then all the possibilities for malicious systems go down to
(virtually) zero within a few hours. (There is a long list of
supporting arguments for this, but that will have to be a separate
discussion).
To argue that there is any possibility of a malevolent AGI emerging, you
have to attack the hinge point of this argument - you have to give a
convincing reason why the first AGI (a) is going to be created by
someone who has genocidal motives, (b) is going to be stable AND
malevolent: something that is probably very hard to achieve.
I believe, overall, that when you look into the details of these
scenarios, what you find is that all the Bad Outcome scenarios involve
assumptions that, when you examine them carefully, are wildly
improbable, or based on pure supposition about technical topics that we
have not resolved yet.
I think we should definitely have structured debate about this.
Richard Loosemore.
> -HRJ
>
> ------------------------------
>
>
> Richard Loosemore wrote:
>
> Your comments below are unfounded, and all the worse for being so
> poisonously phrased. If you read the conversation from the beginning
> you will discover why: Matt initially suggested the idea that an AGI
> might be asked to develop a virus of maximum potential, for purposes
> of testing a security system, and that it might respond by inserting
> an entire AGI system into the virus, since this would give the virus
> its maximum potential. The thrust of my reply was that his entire
idea
> of Matt's made no sense, since the AGI could not be a "general"
> intelligence if it could not see the full implications of the
request.
>
>
> Please feel free to accuse me of gross breaches of rhetorical
> etiquette, but if you do, please make sure first that I really have
> committed the crimes. ;-)
>
>
>
> Richard Loosemore
>
>
------------------------------------------------------------------------
*agi* | Archives <http://www.listbox.com/member/archive/303/=now>
<http://www.listbox.com/member/archive/rss/303/> | Modify
<http://www.listbox.com/member/?&>
Your Subscription [Powered by Listbox] <http://www.listbox.com>
-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com