On Thursday, October 25, 2012 1:10:24 AM UTC-4, Brent wrote:
>
>  On 10/24/2012 9:23 PM, Craig Weinberg wrote: 
>
> Or what if we don't care?  We don't care about slaughtering cattle, which 
>> are pretty smart 
>> as computers go.  We manage not to think about starving children in 
>> Africa, and they *are* 
>> humans.  And we ignore the looming disasters of oil depletion, water 
>> pollution, and global 
>> warming which will beset humans who are our children. 
>>
>
> Sure, yeah I wouldn't expect mainstream society to care, except maybe for 
> some people, I am mainly focused on what seems to be like an astronomically 
> unlikely prospect that we will someday find it possible to make a person 
> out of a program, but won't be able to just make the program itself and no 
> person attached. 
>
>
> Right. John McCarthy (inventor of LISP) worried and wrote about that 
> problem decades ago.  He cautioned that we should not make robots conscious 
> with emotions like humans because then it would be unethical use them like 
> robots.
>

It's arbitrary to think of robots though. It can be anything that 
represents computation to something. An abacus, a card game, anything. 
Otherwise it's prejudice based on form. 

>
>  Especially given that we have never made a computer program that can do 
> anything whatsoever other than reconfigure whatever materials are able to 
> execute the program, I find it implausible that there will be a magical 
> line of code which cannot be executed without an experience happening to 
> someone. 
>
>
> So it's a non-problem for you.  You think that only man-born-of-woman or 
> wetware can be conscious and have qualia.  Or are you concerned that we are 
> inadvertently offending atoms all the time?
>

Everything has qualia, but only humans have human qualia. Animals have 
animal qualia, organisms have biological qualia, etc.
 

>
>  No matter how hard we try, we can never just make a drawing of these 
> functions just to check our math without invoking the power of life and 
> death. It's really silly. It's not even good Sci-Fi, it's just too lame.
>  
>
> I think we can, because although I like Bruno's theory I think the MGA is 
> wrong, or at least incomplete.  I think the simulated intelligence needs a 
> simulated environment, essentially another world, in which to *be* 
> intelligent.  And that's where your chalk board consciousness fails.  It 
> needs to be able to interact within a chalkboard world.  So it's not just a 
> question of going to a low enough level, it's also a question of going to a 
> high enough level.
>

A chalkboard world just involves a larger chalkboard.

Craig
 

>
> Brent
> The person I was when I was 3 years old is dead. He died because
> too much new information was added to his brain.
>          -- Saibal Mitra
>  

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To view this discussion on the web visit 
https://groups.google.com/d/msg/everything-list/-/B-2eXjKhYLYJ.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to 
[email protected].
For more options, visit this group at 
http://groups.google.com/group/everything-list?hl=en.

Reply via email to