On Monday, September 16, 2019 at 10:17:24 PM UTC-6, Brent wrote: > > > > On 9/16/2019 7:49 PM, Alan Grayson wrote: > > > > On Monday, September 16, 2019 at 2:41:26 PM UTC-6, Brent wrote: >> >> >> >> On 9/16/2019 6:07 AM, Alan Grayson wrote: >> > My take on AI; it's no more dangerous than present day computers, >> > because it has no WILL, and can only do what it's told to do. I >> > suppose it could be told to do bad things, and if it has inherent >> > defenses, it can't be stopped, like Gort in The Day the Earth Stood >> > Still. AG >> >> The danger is not so much in AI being told to do bad things, but that in >> doing the good things it was told to do it uses unforseen methods that >> have disasterous consequences. It's like Henry Ford was told to invent >> fast, convenient personal transportation...and created traffic jams and >> global warming. >> >> Brent >> > > One could expect military applications, such as robots replacing human > infantry, their job to kill the enemy. So if their programming had a flaw, > accidental or intentional, these AI infantry could start killing > indiscriminately. > > > Less likely than with human troops who have built in emotions of revenge > and retaliation. > > It would be hard to stop them since they'd come with self defense > functions. AG > > > But we also know a lot more about their internal construction and > functions. We would probably even build in an Achilles heel. > > Brent >
I think you underestimate the evil that men can do, not to mention some bit flips due to cosmic rays that could change their MO's entirely. AG -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/everything-list/c64d04ba-b56e-4e9b-83c2-fb4547d5b754%40googlegroups.com.

