Terren,
Are you ignoring my reply on purpose, or accidentally? If it is on
purpose, that is fine, but if it is by accident then the original
message is replicated below.
-Abram

On Wed, Aug 6, 2008 at 10:24 AM, Abram Demski <[EMAIL PROTECTED]> wrote:
> On Wed, Aug 6, 2008 at 12:04 AM, Terren Suydam <[EMAIL PROTECTED]> wrote:
>>
>> Abram,
>>
>> If that's your response then we don't actually agree.
>
> Sorry, I meant "I agree that Searle's responses are inadequate".
>
>>
>> I agree that the Chinese Room does not disprove strong AI, but I think it is 
>> a valid critique for purely logical or non-grounded approaches. Why do you 
>> think the critique fails on that level?  Anyone else who rejects the Chinese 
>> Room care to explain why?
>
> I explained somewhat in my first reply to this thread. Basically, as I
> understand you, you are saying that the original chinese room does not
> have understanding, but if we modify the argument to connect it up to
> a robot with adequate senses, it could have understanding (if the
> human inside could work fast enough to show it). But, if I am willing
> to grant that such a robot has understanding (despite the human
> controller having no understanding of the data being manipulated),
> then I may very well be willing to grant that the original Chinese
> room has understanding (as I am willing to grant).
>
> I do distrust some philosophy, but other issues I think are very
> important. For example, I am very interested in the foundations of
> mathematics.
>
> -Abram
>
>>
>> (I know this has been discussed ad nauseum, but that should only make it 
>> easier to point to references that clearly demolish the arguments. It should 
>> be noted however that relatively recent advances regarding complexity and 
>> emergence aren't quite as well hashed out with respect to the Chinese Room. 
>> In the document you linked to, mention of emergence didn't come until a 2002 
>> reference attributed to Kurzweil.)
>>
>> If you can't explain your dismissal of the Chinese Room, it only reinforces 
>> my earlier point that some of you who are working on AI aren't doing your 
>> homework with the philosophy. It's ok to reject the Chinese Room, so long as 
>> you have arguments to do it (and if you do, I'm all ears!) But if you don't 
>> think the philosophy is important, then you're more than likely doing Cargo 
>> Cult AI.
>>
>> (http://en.wikipedia.org/wiki/Cargo_cult)
>>
>> Terren
>>
>> --- On Tue, 8/5/08, Abram Demski <[EMAIL PROTECTED]> wrote:
>>
>>> From: Abram Demski <[EMAIL PROTECTED]>
>>> Subject: Re: [agi] Groundless reasoning --> Chinese Room
>>> To: [email protected]
>>> Date: Tuesday, August 5, 2008, 9:49 PM
>>> Terren,
>>> I agree. Searle's responses are inadequate, and the
>>> whole thought
>>> experiment fails to prove his point. I think it also fails
>>> to prove
>>> your point, for the same reason.
>>>
>>> --Abram
>>>
>>
>>
>>
>>
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=108809214-a0d121
Powered by Listbox: http://www.listbox.com

Reply via email to