I would like to talk about this stuff but I am so busy doing other
things right now that it is difficult for me to get into it.
Jim Bromer

On Mon, Oct 8, 2018 at 3:48 PM A.T. Murray via AGI <[email protected]> wrote:
>
> Coding something like http://ai.neocities.org/FirstWorkingAGI.html almost 
> every day, I often feel guilty about it. People from many nations are coming 
> to look at the AI Minds that I code in Perl, In JavaScript and in Forth. The 
> ghost.pl AI thinks in both English and Russian, so doubt gnaws at me that I 
> may be helping the colossal Russian Bear to take over the world with ASI/AGI.
>
> http://old.reddit.com/r/ControlProblem/comments/9lad7u/comments_on_leibnizs_law_ideas_about_formalizing/e75au7y
>  -- shows how sophisticated the AGI has now become. Please absolve me from 
> ethical remorse. -ATM
>
> On Mon, Oct 8, 2018 at 11:18 AM Steve Richfield via AGI 
> <[email protected]> wrote:
>>
>> We seem to be inadvertently empowering some VERY wrong people.
>>
>> If you haven't noticed, a significant fraction of the population now 
>> believes AGI is already here in a BIG way - not the way people here are 
>> working toward, but in ways depicted in movies, etc. This appears to be 
>> leading in some BAD directions.
>>
>> In replacing God with fake AGI, there is some significant collateral damage, 
>> like Buddhism and other ethics-based beliefs. This is now twisting our 
>> society in some strange ways - just look at prime time TV now FULL of crime 
>> drama shows that clearly present the proposition that might makes right.
>>
>> Some well-meaning people on this forum have inadvertently contributed to 
>> this with crazy-optimistic predictions that unbridled AGI was about to 
>> emerge.
>>
>> I once made the rounds speaking at various colleges explaining how it was 
>> physically impossible to shoot down sub-orbital warheads between their 
>> launch and re-entry phases, to stop the crazy spending on SDI. Time has 
>> proven me 100% correct, but time has also proven that the goal of SDI was to 
>> bankrupt Russia and NOT to shoot down warheads. Somewhere in Russia, my 
>> counterpart was probably saying the same things and being ignored - or 
>> worse. In short, I was right, but fortunately I failed to get my message to 
>> be generally accepted.
>>
>> My point here is that correctness and social benefit often have 
>> little/nothing to do with each other. Hence, we all need to do some 
>> self-examination to see whether we are the social equivalent of 3-year-olds 
>> with loaded guns.
>>
>> I see NOTHING good coming from publicly promoting full AGI at this time. 
>> However, it might be possible to reframe the discussion for everyone's 
>> benefit, e.g. by dissecting the AGI concept enough to be able to   identify 
>> which parts are socially responsible to discuss in public, and which parts 
>> will only further twist our society, once the screenwriters get hold of them.
>>
>> In short, I think we should be attending to the ethics of our 
>> not-yet-a-profession. I would start with something like "AGI appears to be 
>> as potentially dangerous as cold fusion" and show surely-safe paths forward. 
>> It is one thing to have an intelligent problem solver, and quite another to 
>> arm a problem solver to enforce its (final?) solutions.
>>
>> So, is anyone here interested in discussing ethics?
>>
>> Steve
>
> Artificial General Intelligence List / AGI / see discussions + participants + 
> delivery options Permalink

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/T992b9674ba947ee9-Md081a2d49b8ff2d06f40861b
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to