So this "hackability" is a technical question about possibility of
closed-source deployment that would provide functional copies of the
system but would prevent users from modifying its goal system. Is it
really important? Source/technology will eventually get away, and from
it any goal system can be forged.

Susceptibility to being tricked into different goal system through
normal communication is not a question of strength of goal system, but
a question of intelligence of system, so that it will be able to
recognize the intent of such forged communication and refuse to act on
it.

On 10/2/07, Mark Waser <[EMAIL PROTECTED]> wrote:
>     Interesting.  I believe that we have a fundamental disagreement.  I
> would argue that the semantics *don't* have to be distributed.  My
> argument/proof would be that I believe that *anything* can be described in
> words -- and that I believe that previous narrow AI are brittle because they
> don't have both a) closure over the terms that they use and b) the ability
> to learn the meaning if *any* new term (traits that I believe that humans
> have -- and I'm not sure at all that the "intelligent" part of humans have
> distributed semantics).  Of course, I'm also pretty sure that my belief is
> in the minority on this list as well.
>
>     I believe that an English system with closure and learning *is* going to
> be a complex system and can be grounded (via the closure and interaction
> with the real world).  And scalable looks less problematic to me with
> symbols than without.
>
>     We may be different enough in (hopefully educated) opinions that this
> e-mail may not allow for a response other than "We shall see" but I would be
> interested, if you would, in hearing more as to why you believe that
> semantics *must* be distributed (though I will immediately concede that it
> will make them less hackable).
>
>         Mark
>
> ----- Original Message -----
> From: "Richard Loosemore" <[EMAIL PROTECTED]>
> To: <[email protected]>
> Sent: Monday, October 01, 2007 8:36 PM
> Subject: **SPAM** Re: [agi] Religion-free technical content
>
>
> > Mark Waser wrote:
> >>> And apart from the global differences between the two types of AGI, it
> >>> would be no good to try to guarantee friendliness using the kind of
> >>> conventional AI system that is Novamente, because inasmuch as general
> >>> goals would be encoded in such a system, they are explicitly coded as
> >>> "statement" which are then interpreted by something else.  To put it
> >>> crudely (and oversimplify slightly) if the goal "Be empathic to the
> >>> needs of human beings" were represented just like that, as some kind of
> >>> proposition, and stored at a particular location, it wouldn't take much
> >>> for a hacker to get inside and change the statement to "Make [hacker's
> >>> name] rich and sacrifice as much of humanity as necessary".  If that
> >>> were to become the AGI's top level goal, we would then be in deep
> >>> doodoo.  In the system I propose, such events could not happen.
> >>
> >> I think that this focuses on the wrong aspect.  It is not the fact that
> >> the goal is explicitly encoded as a statement that is a problem -- it is
> >> the fact that it is in only one place that is dangerous.  My assumption
> >> is that your system basically build it's base constraints from a huge
> >> number of examples and that it is distributed enough that that it would
> >> be difficult if not impossible to maliciously change enough to cause a
> >> problem.  The fact that you're envisioning your system as not having
> >> easy-to-read statements is really orthogonal to your argument and a
> >> system that explicitly codes all of it's constraints as readable
> >> statements but still builds it's base constraints from a huge number of
> >> examples should be virtually as incorruptible as your system (with the
> >> difference being security by obscurity -- which is not a good thing to
> >> rely upon and also means that your system is less comprehensible).
> >
> > Mark,
> >
> > You have put your finger on one aspect of the proposal that came up, in a
> > slightly different way, when Jef Allbright started talking about
> > pragmatics:  the "semantics" of the system.  This is the hardest feature
> > to explain in a short space.
> >
> > I really did consciously mean to have both things, not just distributed
> > representation of the constraints, but also the fact that the semantics of
> > the system is distributed.  This distributed, semi-opaque semantics is
> > what I meant by talking about the propositions not being explicitly
> > encoded, above, and what I also was referring to in my comment to Jef.
> >
> > If the basic knowledge units ("atoms") of the system develop as a result
> > of learning mechanisms + real world interaction (which together make them
> > grounded), then the meaning of any given atom is encoded in the whole web
> > of connections between it and the other atoms, and also by the mechanisms
> > that browse on (/use, /modify) these atoms.  It is not easy to point to an
> > atom and say exactly what it does.
> >
> > This is not an optional part of the framework:  it is crucial.  It is the
> > main reason why the system has some complexity.  It is also the reason why
> > the system can be properly grounded and is scalable (which is what, with
> > an ordinary, conventional AI system, cannot be done because of the complex
> > systems problem).
> >
> > In a sense the system is less comprehensible, but this is only a matter of
> > degree.  I don't think it makes any practical difference to our attempts
> > to govern its behavior.  It is going to be comprehensible enouigh that we
> > can put hooks in for monitoring purposes.
> >
> > The great benefit of this way of doing things is that, once the system has
> > matured to adulthood, it cannot be hacked:  you cannot just write a worm
> > to go around hunting for constraints and modifying them in a regular way
> > (as you might be able to do with ordinary distributed constraints, where
> > the semantics of each individual atom is well defined enough that you can
> > make a clean edit), because if you tried to do this you would destabilize
> > the whole thing and turn it into a gibbering wreck.  It would stop working
> > ... and the effect would be so dramatic that we (and it) could easily set
> > up automatic shutdown mechanisms to intervene in such a case.
> >
> >
> >
> >
> > Richard Loosemore
> >
> >
> >
> > -----
> > This list is sponsored by AGIRI: http://www.agiri.org/email
> > To unsubscribe or change your options, please go to:
> > http://v2.listbox.com/member/?&;
> >
>
>
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?&;
>


-- 
Vladimir Nesov                            mailto:[EMAIL PROTECTED]

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=48736327-159eb6

Reply via email to