> From: [email protected]
> No. AGI is unsolved. I am saying that there are machine learning
> algorithms that build on previous learning. For example, the LZW
> compression algorithm builds its dictionary by extending the words it
> has already learned. Of course this is not AGI because it meets none
> of the requirements of solving language, vision, hearing, robotics,
> art, and predicting human behavior all with human level ability. But
> you don't claim to be trying to solve these problems either.
>
> But you haven't answered my questions. Exactly what will your program
> do? How will you demonstrate that your program "builds on previous
> learning"? What are the tests that you will give it?
Complexity is too great a problem.
1. What I have said is that if I could get my program to work with text based
IO then it could be modified to work with vision hearing robotics etcetera.
The idea that it could predict human behavior with human level ability is from
your definition, not mine.
2. What I have directly implied is that if I can't get it to work with
text-based IO I wouldn't be able to get it to work with vision hearing robotics
etcetera.
However, to continue to get nearer to answering your questions. Since there
are numerous algorithms (narrow AI) that can excel at tasks that even we would
not be able to do, this indicates that if I was able to write a simple AGI
program (which is much less able than the majority of human beings) then there
would be numerous AI things that my program would not be able to even try. So
I would have trouble demonstrating that my Simple AGI was better than other
programs on the kinds of tasks that they excelled in even if the other program
was narrow AI. If simple narrow AI algorithms that input numerical values
from an IO data environment and output values in a numerical range were mixed
together in the right way, we could use the old equivalency argument to argue
that even the simplest narrow AI programs could hypothetically be combined to
produce AGI. This argument suggests that machine learning algorithms are not
much more than a step up from the earlier AI methods. My interest then is in
finding the 'right way' to combine some narrow AI algorithms to produce AGI.
But many people in this group agree that narrow AI programs are not AGI and so
my goal is to build an simple AGI program that could be pitted against other
AGI programs. Even then, another AGI program that included some powerful
narrow methods could presumably beat my program in those particular challenges.
So I started describing conceptual integration and building on previous
learning as a means to begin finding a working definition of what is needed for
an AGI approach. Your responses have give me great hope that I am as far along
in my theories as I thought I was.
I started to answer your question by pointing to a number of ideas. Think
about the problem of an executive function (starting with meta-awareness) .
And look at the integration problem where integration involves more than
subsequent numerical or logical inclusion of a simple narrow kind. The
executive 'function' won't be involved in every detail but it will note some of
the characteristics of the algorithms as they relate to the context of the
application of the algorithm. The idea that an executive function should have
some greater awareness of the context of an application of an algorithm
sometimes seems to me as if it was a mandatory requirement of AGI. The
executive function will not be able to unerringly tell if an applied algorithm
works or not so it will have to rely on other methods like cross-analysis and
attention to structural integration methods. If an analysis leads to a strong
integration of a model that the program had been building than the application
of the algorithm in that situation will be explored further. But obviously,
there has to be a guard against self-confirming artificial delusion. So while a
narrow-AI algorithm could easily defeat my program at the tasks that it was
explicitly designed for, my program, if it works, will be able to place an idea
that it had learned about in a greater context of ideas. It will be able, if
it works, to talk about the idea that it had just learned about. Since my
program has got to be simple, every thing would be at a very primitive level
but it would still be better than a calculator function which does not know
much of anything about the contexts that it is used in. Even this simple
example, describing how an AGI program with some executive meta-awareness that
is aware of some of the characteristics of the context of an application of a
function is different than a calculator function is more subtle than it might
first seem. Depending on what you call awareness or meta-level executive
function one might argue (and wisely so in my opinion) that since certain
calculator functions can produce error remarks in certain situations that even
calculator functions have some meta-level executive artificial 'awareness' of
the application of the function. The only difference is that the calculator
'speaks' and is programmed in different kinds of languages. So then the
meta-awareness and executive functions of an AGI program must be shaped
partially on previous learning so that it can learn to speak sensibly about the
context and results of an application of a narrower AI method in ways that it
hadn't before and in ways that could be different for different deployments of
the native program. Of course, since my program has to be extremely simple,
this 'sensible conversation' is going to be very primitive and open to
criticism. Jim Bromer
> Date: Sat, 3 Aug 2013 19:53:31 -0400
> Subject: Re: [agi] A Very Simple AGI Project
> From: [email protected]
> To: [email protected]
>
> On Sat, Aug 3, 2013 at 6:07 PM, Jim Bromer <[email protected]> wrote:
> >
> > From: [email protected]
> >
> > "Building on previous learning" is kind of vague. Doesn't any machine
> > learning algorithm do that? How will you test your program, measure the
> > results, and compare it to other approaches to solving the same problems?
> > -----------
> >
> > Are you saying that there are machine learning algorithms that constitute
> > working AGI programs?
>
> No. AGI is unsolved. I am saying that there are machine learning
> algorithms that build on previous learning. For example, the LZW
> compression algorithm builds its dictionary by extending the words it
> has already learned. Of course this is not AGI because it meets none
> of the requirements of solving language, vision, hearing, robotics,
> art, and predicting human behavior all with human level ability. But
> you don't claim to be trying to solve these problems either.
>
> But you haven't answered my questions. Exactly what will your program
> do? How will you demonstrate that your program "builds on previous
> learning"? What are the tests that you will give it?
>
> --
> -- Matt Mahoney, [email protected]
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-f5817f28
> Modify Your Subscription: https://www.listbox.com/member/?&
> Powered by Listbox: http://www.listbox.com
-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription:
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com