Stathis Papaioannou wrote:
> Brent Meeker writes:
> 
> 
>>>What I meant was, if a computer program can be associated with 
>>>consciousness, then a rigid and deterministic computer program can also 
>>>be associated with consciousness - leaving aside the question of how 
>>>exactly the association occurs. For example, suppose I have a conversation 
>>>with a putatively conscious computer program as part of a Turing test, and 
>>>the program passes, convincing me and everyone else that it has been 
>>>conscious during the test. Then, I start up the program again with no memory 
>>>saved from the first run, but this time I play it a recording of my voice 
>>>from 
>>>the first test. The program will go through exactly the same resposes as 
>>>during the first run, but this time to an external observer who saw the 
>>>first 
>>>run the program's responses will be no more surprising that my questions 
>>>on the recording of my voice. The program itself won't know what's coming 
>>>and it might even think it is being clever by throwing is some 
>>>"unpredictable" 
>>>answers to prove how free and human-like it really is. I don't think there 
>>>is any 
>>>basis for saying it is conscious during the first run but not during the 
>>>second. I 
>>>also don't think it helps to say that its responses *would* have been 
>>>different 
>>>even on the second run had its input been different, because that is true of 
>>>any record player or automaton.
>>
>>I think it does help; or at least it makes a difference.  I think you 
>>illegitmately 
>>move the boundary between the thing supposed to be conscious (I'd prefer 
>>"intelligent", because I think intelligence requires counterfactuals, but I'm 
>>not 
>>sure about consciousness) and its environment in drawing that conclusion.  
>>The 
>>question is whether the *recording* is conscious.  It has no input.  But then 
>>you say 
>>it has counterfactuals because the output of a *record player* would be 
>>different 
>>with a different input.  One might well say that a record player has 
>>intelligence - 
>>of a very low level.   But a record does not.
> 
> 
> Perhaps there is a difference between intelligence and consciousness. 
> Intelligence 
> must be defined operationally, as you have suggested, which involves the 
> intelligent 
> agent interacting with the environment. A computer hardwired with "input" is 
> not a 
> very useful device from the point of view of an observer, displaying no more 
> intelligence 
> than a film of the screen would. However, useless though it might be, I don't 
> see why 
> the computer should not be conscious with the hardwired input if it is 
> conscious with the 
> same input on a particular run from a variable environment. If the experiment 
> were set 
> up properly, it would be impossible for the computer to know where the input 
> was 
> coming from. Another way to look at it would be to say that intelligence is 
> relative to 
> an environment but consciousness is absolute. This is in keeping with the 
> fact that 
> intelligent behaviour is third person observable but consciousness is not.
> 
> Stathis Papaioannou

Good point. I think I agree.  My functional view of consciousness is that it's 
a 
filter that puts together a story about what's important to remember.  It's 
needed 
for learning and hence for intelligence of higher order - but it's a subsystem 
of 
intelligence in general.

Brent Meeker

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To post to this group, send email to everything-list@googlegroups.com
To unsubscribe from this group, send email to [EMAIL PROTECTED]
For more options, visit this group at 
http://groups.google.com/group/everything-list
-~----------~----~----~----~------~----~------~--~---

Reply via email to