I have no wish to rehash the fairly futile and extremely disruptive
discussion of Loosemore's assertions that occurred on the SL4 mailing
list. I am willing to address the implicit questions/assumptions about my
own position.

Richard Loosemore wrote:
> The contribution of complex systems science is not to send across a
> whole body of plug-and-play theoretical work: they only need to send
> across one idea (an empirical fact), and that is enough. This empirical 
> idea is the notion of the disconnectedness of global from local behavior 
> - what I have called the 'Global-Local Disconnect' and what, roughly 
> speaking, Wolfram calls 'Computational Irreducibility'.

This is only an issue if you're using open-ended selective dynamics on
or in a substrate with softly-constrained, implicitly-constrained or
unconstrained side effects. Nailing that statement down precisely would
take a few more paragraphs of definition, but I'll skip that for now. The
point is that plenty of complex engineered systems, including almost all
existing software systems, don't have this property. The assertion that
it is possible (for humans) to design an AGI with fully explicit and
rigorous side effect control is contraversial and unproven; I'm optimistic
about it, but I'm not sure and I certainly wouldn't call it a fact. What
you failed to do was show that it is impossible, and indeed below you
seem to acknowledge that it may in fact be possible.

The assertion that it is more desirable to build an AGI with strong
structural constraints is more complicated. Eliezer Yudkowsky has
spent hundreds of thousands of words arguing fairly convincingly for
this, and I'm not going to revist that subject here.

>> It is entirely possible to build an AI in such a way that the general
>>  course of its behavior is as reliable as the behavior of an Ideal
>> Gas: can't predict the position and momentum of all its particles,
>> but you sure can predict such overall characteristics as temperature,
>> pressure and volume.

A highly transhuman intelligence could probably do this, though I
suspect it would be very inefficient, partially I expect you'd need
strong passive constraints on the power of local mechanisms (the kind
the brain has in abundance), which will always sacrifice performance
on many tasks compared to unconstrained or intelligently-verified
mechanisms. The chances of humans being able to do this are
pretty remote, much worse than the already not-promising chances
for doing constraint/logic-based FAI. Part of that is due to the fact that
while there are people making theoretical progress on constraint-based
analysis of AGI, all the suggestions for developing the essential theory
for this kind of FAI seem to involve running experiments on highly
dangerous proto-AGI or AGI systems (necessarily built before any
such theory can be developed and verified). Another problem is the
fact that people advocating this kind of approach usually don't
appreciate the difficult of designing a good set of FAI goals in the
first place, nor the difficulty of verifying that an AGI has a precisely
human-like motivational structure if they're going with the dubious
plan of hoping an enhanced-human-equivalent can steer humanity
through the Singularity successfully. Finally the most serious problem
is that an AGI of this type isn't capable of doing safe full-scale self
modification until it has full competence in applying all of this as yet
undeveloped emergent-FAI theory; unlike constraint-based FAI you
don't get any help from the basic substrate and the self-modification
competence doesn't grow with the main AI. Until both the abstract
knowledge of the reliable-emergent-goal-system-design and the
Friendly goal system to use it properly are fully in place (i.e. in all of
your prototypes) you're relying on adversarial methods to prevent
arbitary self-modification, hard takeoff and general bad news.

In short it's ridiculously risky and unlikely to work, orders of magnitude
more so than actively verified FAI on a rational AGI substrate, which is
already extremely difficult and pretty damn risky to develop. Hybrid
approaches (e.g. what Ben's probably envisioning) are almost certainly
better than emergence-based theories (and I use the word theories
loosely there), and I accept that if fully formal FAI turns out to be
impossibly difficult we might have to downgrade to some form of
probabilistic verification.

> The motivational system of some types of AI (the types you would
> classify as tainted by complexity) can be made so reliable that the 
> likelihood of them becoming unfriendly would be similar to the 
> likelihood of the molecules of an Ideal Gas suddenly deciding to split 
> into two groups and head for opposite ends of their container.

Ok, let's see the design specs for one of these systems, along with
some evidence that it's scalable to AGI. Or is this just a personal
hunch?

> And by contrast, the type of system that the Rational/Normative AI 
> community want to build (with logically provable friendliness) is either 
> never going to arrive, or will be as brittle as a house of cards: it 
> will not degrade gracefully.

In many cases it deliberately doesn't degrade gracefully; sane designs
are laced with redundant checks that shut down the system completely
at the first sign of trouble. Any other assertions of brittleness are
probably a layer confusion and misgeneralisation from classic symbolic
AI (possibly even non-AI software systems).

> For that reason, I believe that if/when you do get impatient and decide
> to forgo a definitive proof of friendliness, 

Which I won't. Someone else might of course.

> and push the START button on your AI, you will create something 
> incredibly dangerous.

You appear to have zero understanding of the functional mechanisms involved
in a 'rational/normative' AI system that could actually scale to AGI, and
you have yet to produce any general arguments that tell us something about
the behaviour of such a system (acknowledging that they are designed to
avoid the kind of unexpected global dynamics you are so fond of, and
presumably accepting that this is possible at an unknown potential cost in
flexibility).
 
Michael Wilson
Director of Research and Development
Bitphase AI Ltd - http://www.bitphase.com


-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/[EMAIL PROTECTED]

Reply via email to