On 11/14/06, James Ratcliff wrote:
If the "contents of a knowledge base for AGI will be beyond our ability to
comprehend"  then it is probably not human level AGI, it is something
entirely new, and it will be alien and completely foriegn and unable to
interact with us at all, correct?
  If you mean it will have more knowledge than we do, and do things somewhat
differently, I agree on the point.
  "You can't look inside the box because it's 10^9 bits."
Size is not a acceptable barrier to looking inside.  Wiki, is huge and will
get infineltly huge, yet I can look inside it, and see that "poison ivy
causes rashes" or whatnot.
The AGI will have enourmous complexity, I agree, but you should ALWAYS be
able to look inside it.  Not in the tradional sense of pages of code maybe
or simple set of rules, but the AGI itself HAS to be able to generalize and
tell what it is doing.
  So something like, I see these leafs that look like this, supply picture,
can I pick them up safely, will generate a human readable output that can
itself be debugged. Or asking about the process of doing something, will
generate a possible plan that the AI would follow, and a human could say, no
thats not right, and cause the AI to go back and reconsider with new
possible information.
  We can always look inside the 'logic' of what the AGI is doing, we may not
be able to directly change that ourselves easily.



Doesn't that statement cease to apply as soon as the AGI starts
optimizing it's own code?
If the AGI is redesigning itself it will be changing before our eyes,
faster than we can inspect it.

You must be assuming a strictly controlled development system where
the AGI proposes a change, humans inspect it for a week then tell the
AGI to proceed with that change.
I suspect you will only be able to do that in the very early development stages.


BillK

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to