On 10/10/2018 4:12 PM, Pierz wrote:
It's not intelligent behaviour. There are tons of things (human
artifacts that have been created to automate certain complex
input-output systems) that exhibit complex, intelligent-ish behaviour
that I seriously doubt have any more sentience than a rock, though I'm
open to the possibility of some sentience in rocks. My "method of
determining if something is conscious" is the same as most people who
don't believe their smart phones are having experiences. It's being a
biological organism with a nervous system, though again, I'm agnostic
on organisms like trees. When /you're/ not being a philosopher I bet
that's your real criterion too! You're not worrying about killing your
smartphone when you trash it for the next model.
Of course this is based on a guess, as yours is. My lack of a good
theory of the relationship between matter and mind does not force me
into solipsism because the absence of a test proves nothing about
reality. Things are as they are. All people are conscious, I assume.
Probably all animals. Possibly plants and rocks and stars and atoms,
in some very different way from us. Whatever way it is, it /is/ that
way regardless of whether I can devise a test for it, even in principle.
I generally agree. I like to resort to my intuition pump, the AI Mars
Rover, because I think the present Mars Rovers have a tiny bit of
intelligence and a corresponding tiny bit of consciousness. Their
intelligence is in their navigation, deployment of instruments,
self-monitoring, and reporting to JPL. They make some decisions about
these things, but they don't learn from experience so they probably at
the level of some insects or spiders, except that they have more
language with which they communicate with JPL. But an AI Mars Rover that
was designed to learn from experience would, I think, be made conscious
in some degree. This is because it will need to remember experiences
and recall relevant ones when faced with unusual problems. Solving the
problem by using experience means having a self-model in a simulation to
try to foresee the outcome of different choices. I think that's the
essence of basic consciousness, learning from experience and
self-modeling as part of decisions.
Brent
--
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.