> 1) > Since we humans will be teaching the AGI, and it will be learning by > interacting with humans and reading human literature, it will absorb > something of the human sense of self
I agree that our emodiment, along with our senses, is a primary source of our sense of self. As I look out my window, I see trees and houses. When I look down I see my legs and hand. So There is *me* here, and all that other *stuff* out there. So I *must* be a separate self. An Intelligent person thinking more deeply realizes that what we call the self is made totaly of non-self items. So setting aside metaphysical concepts for the moment, i can even practically see that I am made of stars, and oceans and clouds and dirt and animals and air and conversations etc. etc. etc. So the idea of non-self is not so great a leap..but i digress.. For a computer, the idea of self will be more nebulous for sure. But I am not comforted by the idea that just because they have a more disparate self, that they will in any sense be less harmful. In fact, if its ego equates with its size, it may even be worse than humans!! ;) I'm not convinced that conversing with humans will make it more human, or develop a sense of self. Its all in the code, as I see it. How is the structure set up? Are there links where the idea of self preservation can be developed? The machine does not *really* need to have a sense of self to be dangerous, just to have an algorithm that encodes self protective like actions will be enough to spawn potentially dangerous behavior...IMO I'm not convinced that a sense of self is required to develop an AGI. Of course, a computer that understands that its a computer, and that humans and the rest of the world are "out there", is more useful than one that doesn't understand this most basic of concepts. But this level of understanding does not constitute a *self* that I would be worried about... I think an AGI can exceed humans in many or most ways, yet still have no sense of self or self preservation... In fact, we have computers that do this today, but only in specific domains. I am stating that I think the same is possible for a more general intelligence as well... But I think we all can admit that once an AGI grows and grows and especially if it can self modify, that something tantamount to *self* or conscioussness might emerge... Kevin > ------- > To unsubscribe, change your address, or temporarily deactivate your subscription, > please go to http://v2.listbox.com/member/?[EMAIL PROTECTED] > ------- To unsubscribe, change your address, or temporarily deactivate your subscription, please go to http://v2.listbox.com/member/?[EMAIL PROTECTED]
