Oops, I mean chimpanzees not gorillas. They outperform humans in short term memory tests. https://youtu.be/ravykEih1rE
On Sun, Feb 24, 2019, 9:46 AM Stefan Reich via AGI <agi@agi.topicbox.com> wrote: > > For example, you cannot learn to remember a 20 digit permutation on a > screen and immediately recall it back no matter how much you practice, > which a something a gorilla can do because its DNA is different. > > Gorillas can do that? > > On Sun, 24 Feb 2019 at 14:46, Matt Mahoney <mattmahone...@gmail.com> > wrote: > >> Colin, I think the source of our disagreement is that we have very >> different ideas about what we mean by AGI. To you, AGI is an autonomous >> agent that learns on its own by doing science (experiments). It is >> completely general and can work in any field like a real human. To me, AGI >> means automating anything we might have to otherwise have to pay a human to >> do. >> >> You believe (I think) that AGI is not even possible in conventional >> computers. I tend to agree. First, transistors use too much power, about a >> megawatt for a human brain sized neural network. We might achieve tens of >> kilowatts using neuromorphic computing at the physical limits of >> minituration. >> >> Second, the brain is optimized for reproductive fitness, not universal >> learning, something Legg proved is not even mathematically possible. >> Instead we are born with 10^9 bits of knowledge encoded on our DNA. That is >> half of what we know as adults, and that knowledge took 3 billion years to >> program at the rate of one bit per generation. For example, you cannot >> learn to remember a 20 digit permutation on a screen and immediately recall >> it back no matter how much you practice, which a something a gorilla can do >> because its DNA is different. >> >> Third, we do not even want autonomy. Then we would have to deal with >> human limitations like emotions and the need to sleep, take vacations, and >> get paid. We work around these limitations and train humans to specialize >> in a million different fields because that is how an organization gets work >> done. >> >> I don't expect AGI to look anything like a human. We already automate 99% >> of work using specialized machines that are vastly better at their jobs >> than humans could ever be. We don't want autonomy. We want to be in >> control. We want to asymptomatically approach 100% as the cost of the >> remaining human portion rises at 3-4% per year as it has for centuries. AGI >> is not a robot revolution. AGI is more productivity with less effort using >> machines that can see and understand language, sense but not feel, know >> what we want without wanting, and recognize and predict human emotions >> without having any. >> >> On Sun, Feb 24, 2019, 1:56 AM Colin Hales <col.ha...@gmail.com> wrote: >> >>> Matt: >>> >>> "When you put millions of these specialists together you have AGI." >>> >>> No you don't!!!! Says who? (1) Where's the proof? (2) Where's the >>> principle that suggests it? You have neither of these things. Even if you >>> had both these things you'd still have to build AGI and test it assuming >>> they are false in order to do the science properly. >>> >>> And you do not get to 'define it' this way. >>> >>> This is SCIENCE. You have an artificial version of a natural general >>> intelligence when you've built one and it (yes the AGI itself, not your or >>> anyone else's blessing) AUTONOMOUSLY proves it. Like fire. Flight and a >>> million other things. >>> >>> I can think of 1000 things that such a specialist-narrow-AI-collection >>> doesn't cover (like everything that science does not know, but could find >>> out), and that a natural general intelligence can autonomously learn (find >>> out), but that 'collection of specialist narrow AI' lacks .... along with >>> an ability to _autonomously_ learn it, which is also something natural >>> general intelligence (yes us) can do. And even worse: such specialist >>> collections have ZERO intelligence, not because it doesn't know >>> something, but because it lacks the _autonomous_ bit. A real AGI can know >>> absolutely NOTHING and yet have non-zero intelligence because it includes a >>> means to autonomously find out. Like us. >>> >>> So that collection cannot be an "artificial version of a natural >>> general intelligence". >>> >>> End of story. >>> >>> Real AGI is defined by how it autonomously handles what it doesn't know >>> (its ignorance!), not by what we bestow on it. If we bestow a means for >>> learning X, that too does not make it AGI. A real but artificial version of >>> a natural general intelligence has to _autonomously_ learn how to learn >>> something it does not know, like us. To prove it you first prove it does >>> NOT know it. Then, later, after learning, when it proves it does, >>> autonomously, it gets a stab at the AGI gold medal. >>> >>> Meanwhile? Nothing wrong with what's being done. Powerful. Useful. >>> Impressive. All good. Rah rah rah, carry on. BUT: Its AUTOMATION based on >>> natural general intelligence, not AGI and it has ZERO intellect. >>> >>> [image: AGI.JPG] >>> >>> The entire AGI project is foundered on this basic fact of the science. >>> >>> And all we ever get here is the endless echo chamber of "if only we can >>> program enough computers" (= AUTOMATION) AGI will magically appear. >>> Rubbish. It was rubbish 65 years ago and we've done nothing but prove more >>> completely it is still rubbish. >>> >>> Enough. >>> >>> I know you'll never face it. Forget it. >>> >>> >>> >>> On Sun., 24 Feb. 2019, 11:08 am Matt Mahoney, <mattmahone...@gmail.com> >>> wrote: >>> >>>> OpenCog is one open source effort. But real progress in AI like Google, >>>> Siri, Alexa etc. is not just software. It's hundreds of petabytes of data >>>> from the 4 billion people on the internet and the millions of CPUs needed >>>> to process it. It's not just something you could download and run. >>>> >>>> I realize it's not AGI yet. We are still spending USD $83 trillion >>>> per year for work that machines can't do yet. There are still incompletely >>>> solved problems in vision, language, robotics, art, and modelling human >>>> behavior. That's going to take lots more data and computing power. The >>>> theoretical work is mostly done, although we still lack good models of >>>> humor and music and much of our own DNA. >>>> >>>> If you want to make progress, choose a narrow AI problem. When you put >>>> millions of these specialists together you have AGI. Don't try to do it all >>>> yourself. You can't. >>>> >>>> On Sat, Feb 23, 2019, 12:33 PM Ed Pell <edp...@optonline.net> wrote: >>>> >>>>> All, why is there no open source AGI effort? >>>>> >>>>> Ed Pell >>>>> > > -- > Stefan Reich > BotCompany.de // Java-based operating systems > *Artificial General Intelligence List <https://agi.topicbox.com/latest>* > / AGI / see discussions <https://agi.topicbox.com/groups/agi> + > participants <https://agi.topicbox.com/groups/agi/members> + delivery > options <https://agi.topicbox.com/groups/agi/subscription> Permalink > <https://agi.topicbox.com/groups/agi/T69c9e23ba6b51be9-M7e8c88a665fd613737dfe39d> > ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T69c9e23ba6b51be9-M6eeb8c2df6ee4b86c924fe78 Delivery options: https://agi.topicbox.com/groups/agi/subscription