We have systems that have been evolving for thousands of years to keep
people from doing harmful things; we all contribute to this evolution as we
fight to protect ourselves from each other. Cybercrime has thrown that
establishment for a loop, but it is only a matter of time before evolution
catches up. Once it does, we will be able to apply the same systems towards
keeping both hackers and any rogue AIs they develop from causing serious
harm.


On Mon, May 5, 2014 at 3:25 PM, Mike Archbold <[email protected]> wrote:

> On 5/5/14, Aaron Hosford <[email protected]> wrote:
> > I think the trick lies in multiple redundancies, both for triggering and
> > effecting termination.
> >
> > We should also design in as many mechanisms as possible to avoid the
> > problem in the first place. For example, a very strong negative reward
> > signal for the AGI even considering modifications to certain critical
> zones
> > of its own software or hardware, particularly those that determine the
> > reward levels themselves, in a reinforcement learning-based AGI. (This
> > could be interpreted as an overpowering urge to "stay true to oneself" on
> > the part of the AGI, meaning that it would try to preserve its own
> personal
> > identity.)
> >
>
> These are good ideas IMO.  The real problem is with people, not
> technology.  I mean, had there been no hackers there would have been
> no Target break in, the CEO would still happily have his job.
> Computation is not to blame.
>
>
> >
> > On Mon, May 5, 2014 at 2:51 PM, Mike Archbold <[email protected]> wrote:
> >
> >> There has been a little talk here, but there needs to be a
> >>
> >> SCIENCE OF THE KILL SWITCH.
> >>
> >> There are efforts I know of for mobile kill switch.  Can someone name
> >> a technology that does NOT need a kill switch?
> >>
> >> On 5/5/14, Aaron Hosford <[email protected]> wrote:
> >> > What a silly article. Apparently if you are very intelligent, your
> >> opinion
> >> > matters more than those of experts in the field regardless of your own
> >> > expertise.
> >> >
> >> > I have yet to see a convincing argument that AGI would indeed "take
> >> > over
> >> > the world". It is a technology. Like any technology, we will
> >> incrementally
> >> > improve it in the directions that make the most economic sense. The
> >> > fact
> >> > that the design process (not to mention the adoption of new
> >> > technologies)
> >> > is incremental means that we will have plenty of time to steer clear
> of
> >> the
> >> > design instabilities that would lead to such a debacle, and the fact
> >> > that
> >> > such instabilities are not economically beneficial ensures that even
> if
> >> we
> >> > are short-sighted, we will still have plenty of incentive to avoid
> >> > those
> >> > instabilities.
> >> >
> >> > In other words, why on earth would we design it to do *that*?
> >> >
> >> >
> >> >
> >> > On Sat, May 3, 2014 at 12:40 PM, just camel via AGI <[email protected]>
> >> > wrote:
> >> >
> >> >> One would think that Hawking was way less anthropomorphic? Expecting
> a
> >> >> superintelligent entity to behave like the worst Roman emperor? If
> you
> >> >> have
> >> >> to be anthropomorphic then why not expect them to behave way better
> >> >> than
> >> >> the brightest and most empathic human being? Some of us even stopped
> >> >> eating
> >> >> meat for ethical reasons and I guess that it is safe to assume that
> an
> >> >> advanced AGI will not fight over resources or even atoms in this
> >> universe
> >> >> of abundance.
> >> >>
> >> >> There just is no good reason for an AGI to obsolete humanity against
> >> >> our
> >> >> will. In fact there are so many productive and cooperative options
> >> >> from
> >> >> coexistence to merging to teaching us about the purpose of our
> >> >> existence
> >> >> and helping us to become better beings ...
> >> >>
> >> >> On 05/03/2014 04:57 AM, Alan Grimes via AGI wrote:
> >> >>
> >> >>>
> >> >>> http://guardianlv.com/2014/05/stephen-hawking-tells-truth-
> >> >>> on-ai-perhaps-worst-thing-to-happen-to-humans/
> >> >>>
> >> >>>
> >> >>
> >> >>
> >> >> -------------------------------------------
> >> >> AGI
> >> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> >> RSS Feed:
> >> >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> >> >> Modify Your Subscription: https://www.listbox.com/
> >> >> member/?&
> >> >> Powered by Listbox: http://www.listbox.com
> >> >>
> >> >
> >> >
> >> >
> >> > -------------------------------------------
> >> > AGI
> >> > Archives: https://www.listbox.com/member/archive/303/=now
> >> > RSS Feed:
> >> https://www.listbox.com/member/archive/rss/303/11943661-d9279dae
> >> > Modify Your Subscription:
> >> > https://www.listbox.com/member/?&;
> >> > Powered by Listbox: http://www.listbox.com
> >> >
> >>
> >>
> >> -------------------------------------------
> >> AGI
> >> Archives: https://www.listbox.com/member/archive/303/=now
> >> RSS Feed:
> >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> >> Modify Your Subscription:
> >> https://www.listbox.com/member/?&;
> >> Powered by Listbox: http://www.listbox.com
> >>
> >
> >
> >
> > -------------------------------------------
> > AGI
> > Archives: https://www.listbox.com/member/archive/303/=now
> > RSS Feed:
> https://www.listbox.com/member/archive/rss/303/11943661-d9279dae
> > Modify Your Subscription:
> > https://www.listbox.com/member/?&;
> > Powered by Listbox: http://www.listbox.com
> >
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/23050605-2da819ff
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to