Neither of those objections apply to what I'm saying. I agree that the system as a whole does not have understanding, but that's not what I'm asserting. I making the specific claim that an emergent aspect of the system is the agent of understanding. The brain simulator defense suffers the same problem, that Searle does not adequately address the role of emergence.
Terren --- On Tue, 8/5/08, Abram Demski <[EMAIL PROTECTED]> wrote: > From: Abram Demski <[EMAIL PROTECTED]> > Subject: Re: [agi] Groundless reasoning --> Chinese Room > To: [email protected] > Date: Tuesday, August 5, 2008, 6:07 PM > Terren, > > You and I could agree. But the Chinese Room, as a thought > experiment, > is supposed to refute that. > > The reply you are giving is very similar to the > "Systems Reply": > > http://plato.stanford.edu/entries/chinese-room/#4.1 > > "Searle's response to the Systems Reply is simple: > in principle, the > man can internalize the entire system, memorizing all the > instructions, doing all the calculations in his head. He > could then > leave the room and wander outdoors, perhaps even conversing > in > Chinese. But he still would have no way to attach "any > meaning to the > formal symbols". The man would now be the entire > system, yet he still > would not understand Chinese. For example, he would not > know the > meaning of the Chinese word for hamburger. He still cannot > get > semantics from syntax. (See below the section on Syntax and > Semantics)." > > http://plato.stanford.edu/entries/chinese-room/#4.3 could > also be > taken to apply to your response, but I won't quote that > one. > > Sincerely, > Abram Demski > > > On Tue, Aug 5, 2008 at 1:50 PM, Terren Suydam > <[EMAIL PROTECTED]> wrote: > > > > The Chinese Room argument counters only the assertion > that the computational mechanism that manipulates symbols is > capable of understanding. But in more sophisticated > approaches to AGI, the computational mechanism is not the > agent, it's merely a platform. > > > > Take the OpenCog design. See in particular: > > > > > http://www.opencog.org/wiki/OpenCogPrime:EmergenceOverview > > > > The 'phenomenal self' emerges as a consequence > of interaction with the environment and the continuous > search for explanations of behavior (including its own). It > is this emergent self that is the agent of understanding. > The computational framework that facilitates the interaction > is merely a platform. Nobody would say that the computer > manipulating the symbols has understanding (in accordance > with the Chinese Room), but it is possible that someday > we'd agree that OpenCog the *agent* has achieved it. > > > > Terren > > > > --- On Tue, 8/5/08, Abram Demski > <[EMAIL PROTECTED]> wrote: > >> The original argument was put forward to show that > all AI > >> is > >> impossible, not just symbolic AI. I can see why > someone > >> might take it > >> the other way, but I don't think it works; the > >> complicated instruction > >> books inside the room could implement either a > symbolic AI > >> or a > >> nonsymbolic one. The details would need to be > changed; > >> perhaps you > >> want video input rather than chinese input, and > >> robot-control as > >> output. And we'd need some silly trick like a > timewarp > >> to make the > >> robot move in real time. But I think Searle would > still > >> stick to his > >> guns and say that the person in the room does not > >> comprehend the data > >> he is manipulating (the 1s and 0s from the video > feed), > >> therefore no > >> understanding occurs. > >> > >> So, Terren, in my opinion you should drop the > chinese room > >> in > >> connection with your argument. It simply has too > much > >> historical > >> baggage. > >> > >> -Abram > >> > >> > >> ------------------------------------------- > >> agi > >> Archives: > https://www.listbox.com/member/archive/303/=now > >> RSS Feed: > https://www.listbox.com/member/archive/rss/303/ > >> Modify Your Subscription: > >> https://www.listbox.com/member/?& > >> Powered by Listbox: http://www.listbox.com > > > > > > > > > > > > ------------------------------------------- > > agi > > Archives: > https://www.listbox.com/member/archive/303/=now > > RSS Feed: > https://www.listbox.com/member/archive/rss/303/ > > Modify Your Subscription: > https://www.listbox.com/member/?& > > Powered by Listbox: http://www.listbox.com > > > > > ------------------------------------------- > agi > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/ > Modify Your Subscription: > https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121 Powered by Listbox: http://www.listbox.com
