I would like to talk about this stuff but I am so busy doing other things right now that it is difficult for me to get into it. Jim Bromer
On Mon, Oct 8, 2018 at 3:48 PM A.T. Murray via AGI <[email protected]> wrote: > > Coding something like http://ai.neocities.org/FirstWorkingAGI.html almost > every day, I often feel guilty about it. People from many nations are coming > to look at the AI Minds that I code in Perl, In JavaScript and in Forth. The > ghost.pl AI thinks in both English and Russian, so doubt gnaws at me that I > may be helping the colossal Russian Bear to take over the world with ASI/AGI. > > http://old.reddit.com/r/ControlProblem/comments/9lad7u/comments_on_leibnizs_law_ideas_about_formalizing/e75au7y > -- shows how sophisticated the AGI has now become. Please absolve me from > ethical remorse. -ATM > > On Mon, Oct 8, 2018 at 11:18 AM Steve Richfield via AGI > <[email protected]> wrote: >> >> We seem to be inadvertently empowering some VERY wrong people. >> >> If you haven't noticed, a significant fraction of the population now >> believes AGI is already here in a BIG way - not the way people here are >> working toward, but in ways depicted in movies, etc. This appears to be >> leading in some BAD directions. >> >> In replacing God with fake AGI, there is some significant collateral damage, >> like Buddhism and other ethics-based beliefs. This is now twisting our >> society in some strange ways - just look at prime time TV now FULL of crime >> drama shows that clearly present the proposition that might makes right. >> >> Some well-meaning people on this forum have inadvertently contributed to >> this with crazy-optimistic predictions that unbridled AGI was about to >> emerge. >> >> I once made the rounds speaking at various colleges explaining how it was >> physically impossible to shoot down sub-orbital warheads between their >> launch and re-entry phases, to stop the crazy spending on SDI. Time has >> proven me 100% correct, but time has also proven that the goal of SDI was to >> bankrupt Russia and NOT to shoot down warheads. Somewhere in Russia, my >> counterpart was probably saying the same things and being ignored - or >> worse. In short, I was right, but fortunately I failed to get my message to >> be generally accepted. >> >> My point here is that correctness and social benefit often have >> little/nothing to do with each other. Hence, we all need to do some >> self-examination to see whether we are the social equivalent of 3-year-olds >> with loaded guns. >> >> I see NOTHING good coming from publicly promoting full AGI at this time. >> However, it might be possible to reframe the discussion for everyone's >> benefit, e.g. by dissecting the AGI concept enough to be able to identify >> which parts are socially responsible to discuss in public, and which parts >> will only further twist our society, once the screenwriters get hold of them. >> >> In short, I think we should be attending to the ethics of our >> not-yet-a-profession. I would start with something like "AGI appears to be >> as potentially dangerous as cold fusion" and show surely-safe paths forward. >> It is one thing to have an intelligent problem solver, and quite another to >> arm a problem solver to enforce its (final?) solutions. >> >> So, is anyone here interested in discussing ethics? >> >> Steve > > Artificial General Intelligence List / AGI / see discussions + participants + > delivery options Permalink ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T992b9674ba947ee9-Md081a2d49b8ff2d06f40861b Delivery options: https://agi.topicbox.com/groups/agi/subscription
