Jeez, Will, the point of Artificial General Intelligence is that it can start adapting to an unfamiliar situation and domain BY ITSELF. And your FIRST and only response to the problem you set was to say: "I'll get someone to tell it what to do."

IOW you simply avoided the problem and thought only of cheating. What a solution, or merest idea for a solution, must do is tell me how that intelligence will start adapting by itself - will generalize from its existing skills to cross over domains.

Then, as my answer indicated, it may well have to seek some instructions and advice - especially and almost certainly if it wants to acquire a whole new major skill, as we do, by taking courses etc.

But a general intelligence should be able to adapt to some unfamiliar situations entirely by itself - like perhaps your submersible situation. No guarantee that it will succeed in any given situation, (as there isn't with us), but you should be able to demonstrate its power to adapt sometimes.

In a sense, you should be appalled with yourself that you didn't try to tackle the problem - to produce a "cross-over" idea. But since literally no one else in the field of AGI has the slightest "cross-over" idea - i.e. is actually tackling the problem of AGI, - and the whole culture is one of avoiding the problem, it's to be expected. (You disagree - show me one, just one, cross-over idea anywhere. Everyone will give you a v. detailed,impressive timetable for how long it'll take them to produce such an idea, they just will never produce one. Frankly, they're too scared).


Mike Tintner <[EMAIL PROTECTED]> wrote:

 You must first define its existing skills, then define the new challenge
with some degree of precision - then explain the principles by which it will extend its skills. It's those principles of extension/generalization that are the be-all and end-all, (and NOT btw, as you suggest, any helpful info that the robot will receive - that,sir, is cheating - it has to work these
 things out for itself - although perhaps it could *ask* for info).


Why is that cheating? Would you never give instructions to a child
about what to do? Taking instuctions is something that all
intelligences need to be able to do, but it should be attempted to be
minimised. I'm not saying it should take instructions unquestioningly
either, ideally it should figure out whether the instructions you give
are any use for it.

 Will Pearson




-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com

Reply via email to