On 5/5/14, Aaron Hosford <[email protected]> wrote: > I think the trick lies in multiple redundancies, both for triggering and > effecting termination. > > We should also design in as many mechanisms as possible to avoid the > problem in the first place. For example, a very strong negative reward > signal for the AGI even considering modifications to certain critical zones > of its own software or hardware, particularly those that determine the > reward levels themselves, in a reinforcement learning-based AGI. (This > could be interpreted as an overpowering urge to "stay true to oneself" on > the part of the AGI, meaning that it would try to preserve its own personal > identity.) >
These are good ideas IMO. The real problem is with people, not technology. I mean, had there been no hackers there would have been no Target break in, the CEO would still happily have his job. Computation is not to blame. > > On Mon, May 5, 2014 at 2:51 PM, Mike Archbold <[email protected]> wrote: > >> There has been a little talk here, but there needs to be a >> >> SCIENCE OF THE KILL SWITCH. >> >> There are efforts I know of for mobile kill switch. Can someone name >> a technology that does NOT need a kill switch? >> >> On 5/5/14, Aaron Hosford <[email protected]> wrote: >> > What a silly article. Apparently if you are very intelligent, your >> opinion >> > matters more than those of experts in the field regardless of your own >> > expertise. >> > >> > I have yet to see a convincing argument that AGI would indeed "take >> > over >> > the world". It is a technology. Like any technology, we will >> incrementally >> > improve it in the directions that make the most economic sense. The >> > fact >> > that the design process (not to mention the adoption of new >> > technologies) >> > is incremental means that we will have plenty of time to steer clear of >> the >> > design instabilities that would lead to such a debacle, and the fact >> > that >> > such instabilities are not economically beneficial ensures that even if >> we >> > are short-sighted, we will still have plenty of incentive to avoid >> > those >> > instabilities. >> > >> > In other words, why on earth would we design it to do *that*? >> > >> > >> > >> > On Sat, May 3, 2014 at 12:40 PM, just camel via AGI <[email protected]> >> > wrote: >> > >> >> One would think that Hawking was way less anthropomorphic? Expecting a >> >> superintelligent entity to behave like the worst Roman emperor? If you >> >> have >> >> to be anthropomorphic then why not expect them to behave way better >> >> than >> >> the brightest and most empathic human being? Some of us even stopped >> >> eating >> >> meat for ethical reasons and I guess that it is safe to assume that an >> >> advanced AGI will not fight over resources or even atoms in this >> universe >> >> of abundance. >> >> >> >> There just is no good reason for an AGI to obsolete humanity against >> >> our >> >> will. In fact there are so many productive and cooperative options >> >> from >> >> coexistence to merging to teaching us about the purpose of our >> >> existence >> >> and helping us to become better beings ... >> >> >> >> On 05/03/2014 04:57 AM, Alan Grimes via AGI wrote: >> >> >> >>> >> >>> http://guardianlv.com/2014/05/stephen-hawking-tells-truth- >> >>> on-ai-perhaps-worst-thing-to-happen-to-humans/ >> >>> >> >>> >> >> >> >> >> >> ------------------------------------------- >> >> AGI >> >> Archives: https://www.listbox.com/member/archive/303/=now >> >> RSS Feed: >> >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff >> >> Modify Your Subscription: https://www.listbox.com/ >> >> member/?& >> >> Powered by Listbox: http://www.listbox.com >> >> >> > >> > >> > >> > ------------------------------------------- >> > AGI >> > Archives: https://www.listbox.com/member/archive/303/=now >> > RSS Feed: >> https://www.listbox.com/member/archive/rss/303/11943661-d9279dae >> > Modify Your Subscription: >> > https://www.listbox.com/member/?& >> > Powered by Listbox: http://www.listbox.com >> > >> >> >> ------------------------------------------- >> AGI >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: >> https://www.listbox.com/member/archive/rss/303/23050605-2da819ff >> Modify Your Subscription: >> https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com >> > > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/11943661-d9279dae > Modify Your Subscription: > https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
