Richard,

Good - you hit this one on its head! Continuing...

On 7/22/08, Richard Loosemore <[EMAIL PROTECTED]> wrote:

> Steve Richfield wrote:
>
> THIS is a big question. Remembering that absolutely ANY function can be
>> performed by passing the inputs through a suitable non-linearity, adding
>> them up, and running the results through another suitable non-linearity, it
>> isn't clear what the limitations of "linear" operations are, given suitable
>> "translation" of units or point-of-view. Certainly, all fuzzy logical
>> functions can be performed this way. I even presented a paper at the very
>> 1st NN conference in San Diego, showing that one of the two inhibitory
>> synapses ever to be characterized was precisely what was needed to perform
>> an AND NOT to the logarithms of probabilities of assertions being true,
>> right down to the discontinuity at 1.
>>
>
> Steve,
>
> You are stating a well-known point of view which makes no sense, and which
> has been widely discredited in cognitive science for five decades:


I don't really understand how it is possible to "discredit" a prospective
solution that is not yet known, other than exhibiting people's inability to
arrive at it, e.g. as people have been unable to parse English using
POS-based approaches, given ~40 years to do so.



>  you are stating [one version of] the core of the Behaviorist manifesto.


Close, but not exactly. I believe that there is a common math basis with
some "tweaks" as needed for things that don't "fit the pattern".



> Yes, in principle you could argue that intelligent systems consist only of
> a black box with one gargantuan nonlinear function that maps inputs to
> outputs.


Remembering that there are ~200 different types of neurons, probably some
with different physical structure but the same math, and others with
different math, it would be good to arrive at a full understanding of at
least one of them, and move out from there.



> The trouble is that such a "flat" system is only possible in principle:  it
> would be ridiculously huge, and it gives us no clue about how it becomes
> learned through experience.
>
> So the fact that everything could IN PRINCIPLE be done in this simplistic,
> flat kind of system means nothing.  The devil is in the detals and the
> details are just ridiculous.


Again, I am NOT proposing a single type of building block, but rather a
family with a common mathematical underpinning, plus whatever "special
sauce" these fail to provide.



> One problem is that this idea - this "Hey!! Let's Just Explain It With One
> Great Big Nonlinear Function, Folks!!!!" idea - keeps creeping back into the
> cognitive science-neural nets-artificial intelligence complex.


How about substituting ~200 for One.



> Otherwise sensible people keep accidentally reintoducing it without really
> understanding what they are doing;  without understanding the ramifications
> of this idea.
>
> That is why it is meaningless to say something like "Make that present-day
> PCA. Several people are working on its limitations, and there seems to be
> some reason for hope of much better things to come." There is little reason
> to hope for better things to come (except for the low level mechanisms that
> Derek quite correctly pointed out), because the whole PCA idea is a dead
> end.


I hear that you are quite convinced of this, and if this is true, then I
should also become quite convinced. I just don't yet see how to get there
(mentally burying PCA-like approaches and other similar NN-like views) given
that something like these seem to be working for us.

This seems to be going the way of the discussion on the viability of ad hoc
approaches to AGI we had a couple of months ago, where I asked for the prima
facie case that it should work, and got a bunch of opinions generally to the
effect that people felt that it could work, but couldn't state why they felt
this way. Is that the case here - that you feel that PCA-like approaches
can't work, but you can't make the prima facie case?

> A dead end as a general AGI theory, mark you.  It has its uses.


If I could see just one narrow application where something worked every bit
as well as neurons in people do, then there would be some sort of starting
point. Until then, nothing, not even PCA, would seem to "have its uses".

You have quite rightly moved the level of this discussion up to where it
belongs. Now the challenge seems to be for one of us to "put a stake through
the heart" of the other. You just got my spleen - would you care to take
another shot?!

Steve Richfield



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121
Powered by Listbox: http://www.listbox.com

Reply via email to