Hi Pierz,
On Tue, Sep 2, 2014 at 1:21 PM, Pierz <[email protected]> wrote: > I have to say I find the whole thing amusing. Tegmark even suggested we > should be spending one percent of GDP trying to research this terrible > threat to humanity and wondered why we weren't doing it. Why not? Because, > unlike global warming and nuclear weapons, there is absolutely no sign of > the threat materializing. It's an absolutely theoretical risk based on a > wild extrapolation. To me the whole idea of researching defences against a > future robot attack is like building weapons to defend ourselves against > aliens. So far, the major threat from computers is their stupidity, not > their super-intelligence. It's the risk that they will blindly carry out > some mechanical instruction (think of semi-autonomous military drones) > without any human judgement. Some of you may know the story of the Russian > commander who prevented World War III by overriding protocol when his > systems told him the USSR was under missile attack. The computer systems > f%^*ed up, he used his judgement and saved the world. The risk of computers > will always be their mindless rigidity, not their turning into HAL 9000. > Someone on the thread said something about Google face recognition software > exhibiting behaviour its programmers didn't understand and they hadn't told > it to do. Yeah. My programs do that all the time. It's called a bug. When > software reaches a certain level of complexity, you simply lose track of > what it's doing. Singularity, shmigularity. > So I take it you're not a computationalist? Cheers Telmo. > > > On Tuesday, August 26, 2014 5:05:04 AM UTC+10, Brent wrote: >> >> Bostrom says, "If humanity had been sane and had our act together >> globally, the sensible course of action would be to postpone development of >> superintelligence until we figured out how to do so safely. And then maybe >> wait another generation or two just to make sure that we hadn't overlooked >> some flaw in our reasoning. And then do it -- and reap immense benefit. >> Unfortunately, we do not have the ability to pause." >> >> But maybe he's forgotten the Dark Ages. I think ISIS is working hard to >> produce a pause. >> >> Brent >> >> On 8/25/2014 10:27 AM >> >> Artificial Intelligence May Doom The Human Race Within A Century, Oxford >> Professor >> >> http://www.huffingtonpost.com/2014/08/22/artificial- >> intelligence-oxford_n_5689858.html?ir=Science >> >> >> -- > You received this message because you are subscribed to the Google Groups > "Everything List" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To post to this group, send email to [email protected]. > Visit this group at http://groups.google.com/group/everything-list. > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

