On 1/13/2014 11:16 PM, LizR wrote:
On 14 January 2014 19:42, meekerdb <[email protected] <mailto:[email protected]>>
wrote:
On 1/13/2014 10:18 PM, LizR wrote:
...
So you don't think a discussion of what counts as an AI is a good idea? OK, that's fine
by me (you're the one who wants to discuss it, after all!)
No, I meant I don't think it's a good idea to try to restrict AI to mean a *conscious*
computer. I'm certainly willing to assume it's possible. I don't think a philosophers
zombie is possible.
So let's just assume it's possible, since it's presumably a consequence of Edgar's
theory, that a computer could be conscious,
I don't know that Edgar agrees to that, although it would seem to be a consequence of his
theory. He hasn't said what it would take for a robot to be conscious. Would intelligent
behavior be enough?
Brent
and go back to the original discussion.
Edgar...
We were talking about whether a person can always know if they are in a simulated
reality. Suppose the person is an AI inside a simulation. Would they necessarily know it
was a simulation?
--
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/groups/opt_out.