> The knowledge base has high complexity. You can't debug it. You can examine
> it and edit it but you can't verify its correctness.
While the knowledge base is complex, I disagree with the way in which you're
attempting to use the first sentence. The knowledge base *isn't* so complex
that it causes a truly insoluble problem. The true problem is that the
knowledge base will have a large enough size and will grow and change quickly
enough that you can't maintain 100% control over the contents or even the
integrity of it.
I disagree with the second but believe that it may just be your semantics
because of the third sentence. The question is what we mean by "debug". If
you mean remove all incorrect knowledge, then the answer is obviously "yes, we
can't remove all incorrect knowledge" because odd sequences of observed events
and incomplete knowledge means that globally incorrect knowledge *is* the
correct deduction from experience. On the other hand, we certainly should be
able to debug how the knowledge base operates, make sure that it maintains an
acceptable degree of internal integrity, and responds correctly when it detects
a major integrity problem. The *process* and global behavior of the knowledge
base is what is important and it *can* be debugged. Minor mistakes and errors
are just the cost of being limited in an erratic world.
> An AGI with a correct learning algorithm might still behave badly.
No! An AGI with a correct learning algorithm may, through an odd sequence of
events and incomplete knowledge, come to an incorrect conclusion and take an
action that it would not have taken if it had perfect knowledge -- BUT -- this
is entirely correct behavior, not bad behavior. Calling it bad behavior
dramatically obscures what you are trying to do.
> You can't examine the knowledge base to find out why.
No, no, no, no, NO! If you (or the AI) can't go back through the causal chain
and explain exactly why an action was taken, then you have created an unsafe
AI. A given action depends upon a small part of the knowledge base (which may
then depend upon ever larger sections in an ongoing pyramid) and you can debug
an action and see what lead to an action (that you believe is incorrect but the
AI believes is correct).
> You can't manipulate the knowledge base data to fix it.
Bull. You should be able to correctly come across a piece of incorrect
knowledge that lead to an incorrect decision. You should be able to find the
supporting knowledge structures. If the knowledge is truly incorrect, you
should be able to provide evidence/experiences to the AI that leads it to
correct the incorrect knowledge (or, you could just even just tack the correct
knowledge in the knowledge base, fix it so that it temporarily can't be
altered, and run your integrity repair routines -- which, I contend, any AI
that is going to go anywhere must have).
> At least you can't do these things any better than manipulating the inputs
> and observing the outputs.
No. I can find structures in the knowledge base and alter them. I would
prefer not to. I would strongly prefer that it take the form of a conversation
where I "ask" the AGI what it's reasoning was, where it answers, where I point
out where I believe it's knowledge is incorrect and provide proof, and where it
can then alter its own knowledge base appropriately.
> The reason is that the knowledge base is too complex. In theory you could do
> these things if you lived long enough, but you won't. For practical
> purposes, the AGI knowledge base is a black box.
No. I disagreed with your previous statement and I disagree with the reason.
The knowledge base is not that complex. It is that large. And the AI should
not be a black box *at all*. You should be able to examine any given piece to
any given level of detail at will -- you just can't hold all of it in mind at
once. Yes, that will lead to circumstances where it surprises you -- but we're
not looking for 100% predictability. We're looking for an intelligence with
*bounded* behavior.
> You need to design your goals, learning algorithm, data set and test program
> with this in mind.
Prove to me that the AGI knowledge base is a black box and I will. However,
you have already told me that I "can examine it and edit it" -- so what the
heck do *you* mean by a black box?
> Trying to build transparency into the data structure would be pointless.
> Information theory forbids it.
Bull, information theory does not forbid transparency into the data structure.
Prove this and you would invalidate a huge swath of AGI research. What makes
you say this? I believe that this is the core of your argument and would like
to see *any* sort of evidence/argument to support this grandiose claim.
> I am sure I won't convince you, so maybe you have a different explanation why
> 50 years of building structured knowledge bases has not worked, and what you
> think can be done about it?
Hmmm. Let's see . . . . Codd's paper was published in 1970, so the first
fifteen years were devoted to getting to that point. And SQL wasn't even
commercially available until Oracle was released in 1979, so we're down to only
about half that time. Cyc didn't start until 1984 after Machine Learning
started in the early 1980s. Many people took horrible detours into neural
networks while a lot of the rest were forced to constrain their systems by the
limited computing ability available to them (I remember spending thousands of
dollars per month running biochemistry simulations on a Vax that I was able to
easily run on a PC less than five years later). In the past twenty years,
people have continued to make financially-proven progress in applications like
genome databases.
It looks to me like 50 years of building structured knowledge bases has worked
and that were getting better at it all the time and also that we can do more
and more as space and computing power is getting cheaper and cheaper and
languages and techniques are getting more and more powerful. What hasn't
worked *yet* is self-structuring databases and we're learning more all the time
. . . .
So *prove* to me why information theory forbids transparency of a knowledge
base.
Mark
P.S. Yes, yes, I've seen that Google article before where the author believes
that he "proves" that "Google is currently storing multiple copies of the
entire web in RAM." Numerous debunking articles have also come out including
the facts that Google does not store HTML code, that what is stored *is* stored
in compressed form, and --from Google, itself -- that it does *not* store
sections that do not include "new instances of significant terms". But I can
certainly understand your personal knowledge base deriving the "fact" that
"Google DOES keep the searchable part of the Internet in memory" if that
article is all that you've seen on the topic (though one would have hoped that
an integrity check or a reality check would have prompted further evaluation --
particularly since the article itself mentions that that would require an
unreasonably/impossibly large amount of RAM.)
----- Original Message -----
From: "Matt Mahoney" <[EMAIL PROTECTED]>
To: <[email protected]>
Sent: Wednesday, November 15, 2006 6:41 PM
Subject: Re: [agi] A question on the symbol-system hypothesis
Mark Waser wrote:
>Are you conceding that you can predict the results of a Google
search?
OK, you are right. You can type the same query twice. Or if you live long
enough you can do it the hard way. But you won't.
>Are you now conceding that it is not true that "Models that are simple enough
>to debug are too simple to scale."?
OK, you are right again. Plain text is a simple way to represent knowledge. I
can search and edit terabytes of it.
But this is not the point I wanted to make. I am sure I expressed it badly.
The point is there are two parts to AGI, a learning algorithm and a knowledge
base. The learning algorithm has low complexity. You can debug it, meaning
you can examine the internals to test it and verify it is working the way you
want. The knowledge base has high complexity. You can't debug it. You can
examine it and edit it but you can't verify its correctness.
An AGI with a correct learning algorithm might still behave badly. You can't
examine the knowledge base to find out why. You can't manipulate the knowledge
base data to fix it. At least you can't do these things any better than
manipulating the inputs and observing the outputs. The reason is that the
knowledge base is too complex. In theory you could do these things if you
lived long enough, but you won't. For practical purposes, the AGI knowledge
base is a black box. You need to design your goals, learning algorithm, data
set and test program with this in mind. Trying to build transparency into the
data structure would be pointless. Information theory forbids it. Opacity is
not advantagous or desirable. It is just unavoidable.
I am sure I won't convince you, so maybe you have a different explanation why
50 years of building structured knowledge bases has not worked, and what you
think can be done about it?
And Google DOES keep the searchable part of the Internet in memory
http://blog.topix.net/archives/000011.html
because they have enough hardware to do it.
http://en.wikipedia.org/wiki/Supercomputer#Quasi-supercomputing
-- Matt Mahoney, [EMAIL PROTECTED]
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303