Terren,

You and I could agree. But the Chinese Room, as a thought experiment,
is supposed to refute that.

The reply you are giving is very similar to the "Systems Reply":

http://plato.stanford.edu/entries/chinese-room/#4.1

"Searle's response to the Systems Reply is simple: in principle, the
man can internalize the entire system, memorizing all the
instructions, doing all the calculations in his head. He could then
leave the room and wander outdoors, perhaps even conversing in
Chinese. But he still would have no way to attach "any meaning to the
formal symbols". The man would now be the entire system, yet he still
would not understand Chinese. For example, he would not know the
meaning of the Chinese word for hamburger. He still cannot get
semantics from syntax. (See below the section on Syntax and
Semantics)."

http://plato.stanford.edu/entries/chinese-room/#4.3 could also be
taken to apply to your response, but I won't quote that one.

Sincerely,
Abram Demski


On Tue, Aug 5, 2008 at 1:50 PM, Terren Suydam <[EMAIL PROTECTED]> wrote:
>
> The Chinese Room argument counters only the assertion that the computational 
> mechanism that manipulates symbols is capable of understanding. But in more 
> sophisticated approaches to AGI, the computational mechanism is not the 
> agent, it's merely a platform.
>
> Take the OpenCog design. See in particular:
>
> http://www.opencog.org/wiki/OpenCogPrime:EmergenceOverview
>
> The 'phenomenal self' emerges as a consequence of interaction with the 
> environment and the continuous search for explanations of behavior (including 
> its own). It is this emergent self that is the agent of understanding. The 
> computational framework that facilitates the interaction is merely a 
> platform. Nobody would say that the computer manipulating the symbols has 
> understanding (in accordance with the Chinese Room), but it is possible that 
> someday we'd agree that OpenCog the *agent* has achieved it.
>
> Terren
>
> --- On Tue, 8/5/08, Abram Demski <[EMAIL PROTECTED]> wrote:
>> The original argument was put forward to show that all AI
>> is
>> impossible, not just symbolic AI. I can see why someone
>> might take it
>> the other way, but I don't think it works; the
>> complicated instruction
>> books inside the room could implement either a symbolic AI
>> or a
>> nonsymbolic one. The details would need to be changed;
>> perhaps you
>> want video input rather than chinese input, and
>> robot-control as
>> output. And we'd need some silly trick like a timewarp
>> to make the
>> robot move in real time. But I think Searle would still
>> stick to his
>> guns and say that the person in the room does not
>> comprehend the data
>> he is manipulating (the 1s and 0s from the video feed),
>> therefore no
>> understanding occurs.
>>
>> So, Terren, in my opinion you should drop the chinese room
>> in
>> connection with your argument. It simply has too much
>> historical
>> baggage.
>>
>> -Abram
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription:
>> https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>
>
>
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121
Powered by Listbox: http://www.listbox.com

Reply via email to