Jeez, Will, the point of Artificial General Intelligence is that it can
start adapting to an unfamiliar situation and domain BY ITSELF. And your
FIRST and only response to the problem you set was to say: "I'll get someone
to tell it what to do."
IOW you simply avoided the problem and thought only of cheating. What a
solution, or merest idea for a solution, must do is tell me how that
intelligence will start adapting by itself - will generalize from its
existing skills to cross over domains.
Then, as my answer indicated, it may well have to seek some instructions and
advice - especially and almost certainly if it wants to acquire a whole new
major skill, as we do, by taking courses etc.
But a general intelligence should be able to adapt to some unfamiliar
situations entirely by itself - like perhaps your submersible situation. No
guarantee that it will succeed in any given situation, (as there isn't with
us), but you should be able to demonstrate its power to adapt sometimes.
In a sense, you should be appalled with yourself that you didn't try to
tackle the problem - to produce a "cross-over" idea. But since literally no
one else in the field of AGI has the slightest "cross-over" idea - i.e. is
actually tackling the problem of AGI, - and the whole culture is one of
avoiding the problem, it's to be expected. (You disagree - show me one, just
one, cross-over idea anywhere. Everyone will give you a v.
detailed,impressive timetable for how long it'll take them to produce such
an idea, they just will never produce one. Frankly, they're too scared).
Mike Tintner <[EMAIL PROTECTED]> wrote:
You must first define its existing skills, then define the new challenge
with some degree of precision - then explain the principles by which it
will
extend its skills. It's those principles of extension/generalization
that
are the be-all and end-all, (and NOT btw, as you suggest, any helpful
info
that the robot will receive - that,sir, is cheating - it has to work
these
things out for itself - although perhaps it could *ask* for info).
Why is that cheating? Would you never give instructions to a child
about what to do? Taking instuctions is something that all
intelligences need to be able to do, but it should be attempted to be
minimised. I'm not saying it should take instructions unquestioningly
either, ideally it should figure out whether the instructions you give
are any use for it.
Will Pearson
-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com