Boris, I would like to introduce your AGI to a magician friend of mine.
Steve On Thu, Mar 7, 2019, 12:05 Boris Kazachenko <[email protected]> wrote: > "But why would you think that AGI would not hallucinate?" > > Your "AGI" may hallucinate, because it is designed to feed on that > incoherent second-hand natural-language data. > Mine won't, it is designed to be integral and self-sufficient. It will > believe what it sees, not what a bunch of nuts on the net say. > > > On Tue, Mar 5, 2019 at 2:52 PM Linas Vepstas <[email protected]> > wrote: > >> >> >> On Tue, Mar 5, 2019 at 12:37 PM Matt Mahoney <[email protected]> >> wrote: >> >>> Steve, good luck ending the political debate over climate change. But >>> you have a few obstacles. >>> >>> 1. Overwhelming evidence does not end political debate. Just ask the >>> creationists, anti-vaxxers, moon landing hoaxers, and 9/11 conspiracy >>> theorists. The whole purpose of the flat earth society is not to >>> convince you that the earth is flat, but to show that what you think >>> are logical, sound, and obvious arguments supported by undisputed >>> facts are actually useless. >>> >> >> More precisely: computers have lowered the cost of publishing so low, >> that anyone can publish: you no longer run a gauntlet of editors, printers >> and proof-readers who tell you that your ideas are stupid. Facebook and >> twitter have taken this to a new level. As a result, we can now all hear >> each-others brains thinking, and, it turns out, they are incoherent, >> contradictory and insane. Worse: this high-connectivity, high-bandwidth >> (youtube) low-latency interconnect allows for the spread and amplification >> of "memes" disconnected from any basis in reality. >> >> In short: social media has attached our individual, singular brains into >> a big "global brain", and we are hearing that global brain think, and it is >> hallucinating a lot of the time. Old-school social sciences have already >> studied this: propaganda, (Hitler studied propaganda), cults, brainwashing, >> Stockholm syndrome, and also plenty of less harmful things: everything from >> pop music and Puerto-Rican low-rider automotive clubs. Uplifting things, >> too: from scientists to medical doctors to humanitarian activists. >> >> If you want a preview of what a mildly super-human intelligence viz AGI >> might think, then the hallucinatory beliefs of various memetic tribes is a >> good sampling. The memeplex of creationists, anti-vaxxers, moon landing >> hoaxers, and 9/11 conspiracy >> theorists is sufficiently self-consistent to be stable, and more: >> sufficient to be invasive, and occupy the thought-space, the noosphere of >> many human brains. It spreads. >> >> Perhaps you think that an AGI will be purely "rational" (whatever that >> means!) and that no AGI could ever be a creationist anti-vaxxer >> moon-landing denialist. But why would you think that AGI would not >> hallucinate? >> >> Until very recently, Reality, i.e. the universe, entrained thinking minds >> in such a way that you starve and die, if you cannot think clearly enough >> to obtain food and procreate. Squirrels who neglect nuts die sooner, rather >> than later. Our modern economy is sufficiently robust that you can >> hallucinate all day long, or watch soap operas on TV, or do whatever it is >> you do, and mostly not starve to death. Mostly; we seem to have a problem >> with Amazon employees in LA and SF. But whatever. >> >> If you want a safe, non-existential-threatening AGI in the future, >> attempting to understand and control the root causes of hallucinatory >> thinking today is a good place to start. >> >> -- Linas >> -- >> cassette tapes - analog TV - film cameras - you >> > *Artificial General Intelligence List <https://agi.topicbox.com/latest>* > / AGI / see discussions <https://agi.topicbox.com/groups/agi> + > participants <https://agi.topicbox.com/groups/agi/members> + delivery > options <https://agi.topicbox.com/groups/agi/subscription> Permalink > <https://agi.topicbox.com/groups/agi/Tbefabf50a1da4070-M4c500c48685c384665cd0508> > ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Tbefabf50a1da4070-M766237dc4983a45e6680f192 Delivery options: https://agi.topicbox.com/groups/agi/subscription
