Kevin wrote:
!!!!!!!!!!!!!!!
*************
In practice, it seems that an AGI is likely to have an
"owner" or a handful of them, who will have the kind of power you
describe. For instance, if my team should succeed in creating a true
Novamente AGI, then even if others participate in teaching the system, we will
have overriding power to make the changes we want. This goes along with
the fact that artificial minds are not initially going to be given any "legal
rights" in our society (whereas children have some legal rights, though not as
many as adults).
************
Would this overriding occur because the person carries
more weight with Novamente, or would they need to go in and altar the
structure\links\nodes directly to affect the change?
!!!!!!!!!!!!!
Either case could occur. In Novamente, it is possible to assign
default "confidence levels" to information sources, so one could actually tell
the system to assign more confidence to information from certain
individuals. However, there is a lot of flexibility in the design, so
the system could definitely evolve into a configuration where it worked around
these default confidence levels and decided NOT to assign more confidence to
what its teachers told it.
"Going in and altering the structure/links/nodes directly" isn't always
difficult, it may just mean loading a script containing some new (or
reweighted) nodes and links.
!!!!!!!!!!!
*********************************
At least two
questions come up then, right?
1) Depending on the AGI architecture, enforcing one's opinion on the AGI
may be very easy or very difficult. [In Novamente, I guess it will be
"moderately difficult"]
***********************************
That's the crux of the matter
isn't it? Wouldn't it be easy to enforce an opinion while Novamente is
in its formative stages, versus when a large foundation of knowledge is in
place?
!!!!!!!!!!
Yes, that's correct.
**********************************
!!!!!!
Suppose I am overtaken by greed, and I happen to get my hands
on a baby Novamente. I teach it that it should listen to me above
others. I also teach it that it is very desirable for me to have alot of
money. Novamente begins to form goal nodes geared towards fulfilling my
desire for wealth. I direct it to spread itself on the internet, and
determine ways to make me money, preferably without detection. Perhaps
it could manipulate markets, I don't know. Or perhaps it could crack
into electronic accounts and transfer the money to yours truly.
What's to stop\prevent this? In a real sci fi scenario, perhaps for
your next book, could we have NOvamentes "fighting" Novamente's?
!!!!!!!!
There is nothing in the Novamente architecture preventing this kind of
unfortunate occurence. This has to do with the particular system of
goals, beliefs and habits inside a given Novamente system, rather than with
the AI architecture itself.
!!!!!
This all goes to my concern regarding morality. I know you
resist the idea of hard coding morality into the Novamentes for various
reasons. Perhaps as an alternative, the first Novamente could be trained
over a period of time with a strong basis of moral rules(not encoded, but
trained). Then any new Novamentes would be trained by that Novamente
before being released to the public domain, making it nearly impossible for
the new Novamentes to be taught otherwise.
!!!!!!
This is something close to what we have planned.
Several others have asked me about this, and I have promised to write a
systematic (probably brief) document on Novamente Friendliness sometime in
early 2003, shortly after finishing my work on the current draft of the
Novamente book.
!!!
I know some of this stuff is a bit out there, but shouldn't we
be considering this stuff now instead of later??
!!!
It definitely needs to be thought about very hard before Novamente
reaches chimp-level intelligence. And in fact I *have* thought about it
pretty hard, though I haven't written up my thoughts much (as I've prioritized
writing up the actual design, which is taking longer than I'd hoped as it's so
damn big...).
Right now Novamente is just a software core plus a bunch of
modules-being-tested-but-not-yet-integrated, running on top of the core.
So we have a whole bunch of coding and (mostly) testing and tuning to do
before we have a system with animal-level intelligence. Admittedly,
though, if our design is right, the transition from animal-level to
human-level intelligence will be a matter of getting more machines and doing
more parameter-tuning, it won't require introduction of significant new code
or ideas.
Having said that I've thought about and will write about it, however, I
have a big caveat...
My strong feeling is that any theorizing we do about AI morality in
advance, is probably going to go out the window once we have a chimp-level AGI
to experiment with. the important thing is that we go into that phase of
experimentation with the right attitude -- with a realization that training
the system for morality is as important as training it for intelligence -- and
with a careful approach.
-- Ben