Ben Conceptually, I like where you are going with this. Your team's work with human-language-based robotic communication is astounding.
I think your idea of a universal attractor has merit. I suppose, in the end, when matter exists, it generates an elcetro-magnetic field. In a genetic sense, the flux of such a field would act as an open and closed-loop communications network. In this sense, the relevant data, information, and a relative perspective of knowledge, would all be packaged within relative, genomic code. In other words, we are imagining a relative system of relative systems from which reality would functionally emerge. Given my systems methodology, what remains to be done in order to visualize a model of human-like machine reasoning, is to be able to link your "attractor" value to the information, from which it should become possible to systematically emerge any informational concept at any level of abstraction within any, dimension of reasoning. The genetics of resultant information would in theory make forward and backchaining possible, and much more. The completeness schema of functional, attractor values seems to be a critical machine-reasoning component to pursue. It would probably also assume the role of a priority systems constraint. I've been doing much thinking about this as a next-step for my own research. I think you've got this. Keep up the great work. Rob ________________________________ From: Ben Goertzel <[email protected]> Sent: 27 July 2017 04:57 AM To: AGI Subject: Re: [agi] Neuroscience-Inspired AI Well I would say that none of the work done at Deep Mind and also none of the ideas in Demis etc.'s paper address the questions I raised in this paper http://ieeexplore.ieee.org/document/6889662/ How might the brain represent complex symbolic knowledge? - IEEE Xplore Document<http://ieeexplore.ieee.org/document/6889662/> ieeexplore.ieee.org A novel category of theories is proposed, providing a potential explanation for the representation of complex knowledge in the human (and, more generally, (sorry for the paywall ... use sci-hub.cc ...) So there is no real plan for how to achieve abstract symbolic reasoning as needed for human level general intelligence within a purely formal-NN type approach Obviously in opencog we are taking more of a symbolic-neural approach so we don't have issues with abstraction Also if you look at the recent Markram et al paper on algebraic topology and mesoscopic brain structure, there is nothing in the Hassabis etc. universe that seems to address how such structures would be learned or would emerge But sure in a big-picture historical sense the progress happening these days on "narrow AI verging toward AGI" and on "making complex cognitive architectures finally do stuff" is super exciting. We are on the verge of multiple breakthroughs within the next few years. Woo hoo !! - -Ben On Thu, Jul 27, 2017 at 5:55 AM, EdFromNH . <[email protected]<mailto:[email protected]>> wrote: About the above linked Hassabis paper, Ben said, "It's sort of a high level inspirational paper... it does lay down pretty clearly what sort of thinking and approach Deep Mind is likely to be taking in the next years ... there are no big surprises here though as this has been Demis's approach, bias and interest all along, right?" >From my knowledge of several articles and videos by, or about, Hassabis -- I >totally agree. But I am a little less ho-hum than Ben, perhaps because I'm >not as up on the current state of AGI as Ben. Reading Hassabis's paper makes me bullish about how close we are to powerful, if not fully human-level AGI, within 5 years. Why? Because all of the unsolved challenges Hassabis discusses seem like they could be easily solved if enough engineering and programming talent was thrown at them. I feel like I could relatively easily -- within a few months -- weave plausible high level architectural descriptions for solving all of these problems, as, presumably, people like Demis and Ben could do even better. (Perhaps that is why Ben is so ho-hum about the paper.) With the money that's being thrown into AGI, and the much greater ease of doing cognitive architectural experiments made possible with Neural Turing Machines -- which allow programmable, modular plug-and-play with pre-designed and pre-trained neural net modules -- the world is going to get weird fast. Tell me why I am wrong. On Sun, Jul 23, 2017 at 8:29 PM, Ed Pell <[email protected]<mailto:[email protected]>> wrote: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5467749/ On 7/23/2017 4:18 PM, Giacomo Spigler wrote: An Approximation of the Error Backpropagation Algorithm in a Predictive Coding Network with Local Hebbian Synaptic Plasticity ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/8630185-a57a74e1 Modify Your Subscription: https://www.listbox.com/member/?& Powered by Listbox: http://www.listbox.com AGI | Archives<https://www.listbox.com/member/archive/303/=now> [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc] <https://www.listbox.com/member/archive/rss/303/19237892-5029d625> | Modify<https://www.listbox.com/member/?&> Your Subscription [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] <http://www.listbox.com> -- Ben Goertzel, PhD http://goertzel.org "I am God! I am nothing, I'm play, I am freedom, I am life. I am the boundary, I am the peak." -- Alexander Scriabin AGI | Archives<https://www.listbox.com/member/archive/303/=now> [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc] <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> | Modify<https://www.listbox.com/member/?&> Your Subscription [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] <http://www.listbox.com> ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
