On 10/10/2018 4:12 PM, Pierz wrote:
It's not intelligent behaviour. There are tons of things (human artifacts that have been created to automate certain complex input-output systems) that exhibit complex, intelligent-ish behaviour that I seriously doubt have any more sentience than a rock, though I'm open to the possibility of some sentience in rocks. My "method of determining if something is conscious" is the same as most people who don't believe their smart phones are having experiences. It's being a biological organism with a nervous system, though again, I'm agnostic on organisms like trees. When /you're/ not being a philosopher I bet that's your real criterion too! You're not worrying about killing your smartphone when you trash it for the next model.

Of course this is based on a guess, as yours is. My lack of a good theory of the relationship between matter and mind does not force me into solipsism because the absence of a test proves nothing about reality. Things are as they are. All people are conscious, I assume. Probably all animals. Possibly plants and rocks and stars and atoms, in some very different way from us. Whatever way it is, it /is/ that way regardless of whether I can devise a test for it, even in principle.

I generally agree.  I like to resort to my intuition pump, the AI Mars Rover, because I think the present Mars Rovers have a tiny bit of intelligence and a corresponding tiny bit of consciousness. Their intelligence is in their navigation, deployment of instruments, self-monitoring, and reporting to JPL.  They make some decisions about these things, but they don't learn from experience so they probably at the level of some insects or spiders, except that they have more language with which they communicate with JPL. But an AI Mars Rover that was designed to learn from experience would, I think, be made conscious in some degree.  This is because it will need to remember experiences and recall relevant ones when faced with unusual problems.  Solving the problem by using experience means having a self-model in a simulation to try to foresee the outcome of different choices.  I think that's the essence of basic consciousness, learning from experience and self-modeling as part of decisions.

Brent

--
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to