These are exactly the points I am trying to make. It will take time for us
to create these systems, and there will be many iterations of design before
they are even as intelligent as ourselves. The risks of this technology
threatening our society will be minimal, because we will weed out the
problems before it becomes smarter than us. Likewise, the technology is not
going to be magically morally superior to us; we will have to *design *it
to be so. We will do just that, but there is nothing special about
intelligence that will cause it to happen on its own, without our design
efforts focused in that direction. It will require us to invest that
engineering effort, and we will invest it as a result. You made no mention
of the design process, and so you came across (to me) as expecting the AGI
to develop moral superiority as an artifact of its intelligence. If that's
not what you intended, then I apologize for the misunderstanding.

Universal notions of right and wrong based on entropy are another matter,
however. I do not see how you can tie any particular moral (as opposed to
economic) merit to lower entropy. For example, the system might choose to
kill someone in order to free up valuable resources for use by someone who
is more economically productive. Where, in the laws of physics, is the
immorality of such an act encoded?





On Mon, May 12, 2014 at 10:43 AM, just camel via AGI <[email protected]>wrote:

> But then it's also not yet superintelligent and can not yet
> destroy/obsolete our species? Just like a person with down syndrome
> probably can't destroy/obsolete it.
>
>
> On 05/12/2014 03:54 PM, Aaron Hosford wrote:
>
>> Bugs happen. The truth is, the first few versions of this technology are
>> going to suck -- until we improve it. This happens with every new
>> technology.
>>
> It does not need the "human notion" of "right and wrong". There are
> absolute/universal notions of right and wrong. Lower entropy states are
> more profitable and thus "right".
>
> Also why do you imply that something vastly more intelligent than us and
> something which grew within our society would not understand our notions of
> right and wrong? That makes no sense. We won't grab into the Yudkowskian
> "Mindspace" and pick out some random fully fledged agent with predefined
> properties. Whatever AGI system we are talking about will need to evolved
> based on our knowledge pool and of course it will be confronted with our
> notions of right and wrong.
>
>
>
>> On 05/12/2014 03:54 PM, Aaron Hosford wrote:
>>
>>> AGIs won't know, understand, or (especially) care about the human
>>> notions of right and wrong, good and evil, unless we design it to do so.
>>>
>>
>>
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> Modify Your Subscription: https://www.listbox.com/
> member/?&
> Powered by Listbox: http://www.listbox.com
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to