Hi Colin You seem to be following a similar process to AI as to what was used to develop the first, nuclear bomb - various approaches were used coupled with great experimentation.
Semantically, your inclusion of the term "emergent" in your last message undersores this approach for me. I'd like to dwell on its relevance for a few seconds. Emergence is regarded as the basis for complex-systems engineering (Checkland). Further, Checkland asserted how the debate between complex and simple systems would probably give rise to what is regarded as systems thinking. This is ancient stuff I'm repeating only to stress the importance of its credibility. Thus, on the theoretically basis alone, your experimental approach could be deemed to be sound. Narrow AI, broad AI, AGI? All peas in the same pod of complex-systems thinking. The fundamentals still have no significant incentive to change. Personally, I would value such an experimental approach on the basis of rethinking the whole idea of developing AI. How else was the sound-barrier broken? In addition, if one followed the emerging trend in recent, adaptively-autonomous technologies, one would be hard pressed to write off your approach. Just one theoretically-moot point if I may, albeit a semantic one? Any institutionalised process effectively is a program of code. As an extension, any reduced process - as a procedural implementation - on a computer would become a computerized program. Hence, I suppose, your search for a generic algorithmic platform. In the sense of systemically, as soon as you'd link the "stochastic" environment to a computer chip in any way it should emerge as a form of computer program. Whilst one understands the need for research to be highly focussed on its objectives, one must still have a design framework that would not unduly restrict any design in a short-sighted Heisenbergian-Einstein debate. I would assume then that you do have a quantum-based design framework you're working from. If not though, this particular, organic approach would sooner or later come up against the eco-systemic realities of highly-abstracted implementation. This then, mainly due to the lack of navigational competency in the R&D framework to consistently and reliably perform adaptive integration. If it cannot be measured somehow, it cannot be reliably tested and I'm by no means suggesting this to be the case with your experiment. Mine are just thoughts on the interesting topic at hand. One day, when bootstrapping does occur, you'd be wanting to debug though. If only purely mathematical, then purely computational? Maybe that was how computer science emerged. Good luck with the experiment. Rob To: [email protected] From: [email protected] Subject: RE: [agi] Re: Starting to Define Algorithms that are More Powerfulthan Narrow AI Date: Wed, 6 May 2015 10:01:33 +1000 Hi, Rather busy... Having trouble devoting time here. Jim.... You ask if I am making some kind of electric circuit. Basically yes. Except it's physical instantiation is important. Materials in space. I know you won't get why that is important. That's ok for now. Just accept that it's like that for the same reason the brain is like that. What it isn't is an 'Equivalent circuit' in the traditional sense of voltage/current replication. It is designed to produce functionally equivalent action potential-style signaling AND the brain-style field system that actually expresses the voltages. The hardware will (in the field version) express an EEG and MEG like brains. Having said that I am currently designing a version that doesn't express the fields but allows their addition later...knowing what performance degradation results (it will be narrow-AI not AGI). Call it a causality mirror with a faked image in it. It is deeply self modifying. The circuits literally rewire themselves. Circuit loops duplicate/diverge and switch out/off. It accounts for the process of brain development as a kind of learning. I.e. I don't even have to design the 'brain'. It will self configure based on being in the world. Because it's not using neurons it won't automatically mimic brains in structure. I have no idea what a brain will look like. Physically its a crystalline rock. No actual material growth. Functionally it will stabilize in ways I can't know except by experiment. It means that it must be permanently juvenile.. Overexpressed neurons and overexpressed synapses culled back. Lots of wastage. But so what? Not one line of software anywhere. Any 'algorithm' it has is in the adaptation mechanisms. But they are in hardware. The state of the chip's self configuration is the only actual data involved. Yet, when you look at it there will be deep regularities in its behaviour. You could write them down. However they are all emergent. You know what the hardest part of this is? ... Giving it goals. A reason to bother. A reason for it to sustain the quasi-stable resonances that signify its functioning. I have to think of something akin to homeostasis to keep it going! ROBEOSTASIS. You know what might happen? It possibly self-sustain without human intervention or some kind of hardwiring until the fields are added. Unsure. Answering that is an experimental goal. Steve seems to be deeply inside homeostatic concerns. So that's good. I'm not here to justify anything. Experimental proof will speak for me. And if I can't get the version with and without the fields to be different in predicted ways then I will grovel at the feet of the great god computationalism. Not before.😊 I think the approach is a reversion to 'natural cybernetics' that had a brief life in the 1950s and then was lost in a tsunami called computer science. I bring it back for an upgrade. Notice that AGI failure started the moment cybernetics stopped. The actual science of artificial intelligence stopped then, too...IMO. Enough poking the bear. Gotta get back to it. I really appreciate the interest in this 'adaptive control' approach. Cheers Colin From: Jim Bromer Sent: ‎4/‎05/‎2015 12:42 AM To: AGI Subject: Re: [agi] Re: Starting to Define Algorithms that are More Powerfulthan Narrow AI I thought the ideas are interesting and Colin's description was more readable than usual but the arguments supporting the method weren't very powerful. I am curious about how Colin is implementing the method. Could you give me a little more about that? Are you designing some kind of electrical circuit? What I was trying to say in this thread is that you have to supply a little more insight about why you think that the methods that you are designing and will be implementing would rise above being 'narrow ai'. For instance, Colin's honest report on how far he has actually gotten so far sounds like it is on par with simple narrow AI. As I reread your messages I keep finding a little more in it. But back to my point. Since I can rough out the algorithms that I would use as if they were abstractions, or as if they could exist within an abstract world, it would seem that I should be able to conduct simple tests to show that they could diversify in some way that is: 1. at least better than narrow ai, and 2. useful in some way. So perhaps I should add that. I would say, for example, that artificial neural networks would pass this kind of test. However, the criticism then is, ironically given our use of the narrow ai term, that they lack efficient means to focus and they cannot be efficiently used as componential objects. So, can you guys define some abstract or simple tests that could show that your ideas would become able to adapt to the more complicated demands of actual tests? The value of the simple test is that once you can get your algorithms to pass the first test you might come up with ways to design a slightly more aggressive test. So if I could test my ideas to,say, try to learn to recognize some simple classifications then I might try to see if I can get it to try to get it to learn to utilize systems of classifications effectively and efficiently (without redesigning the program only for that specific kind of test.) So then I would have to design some other kind of test to make sure that it is somewhat general. Jim Bromer On Sun, May 3, 2015 at 3:25 AM, Colin Hales <[email protected]> wrote: > > >> On Sat, May 2, 2015 at 2:50 AM, Steve Richfield <[email protected]> >> wrote: >>> >>> Jim, >>> >>> Again, I think I see the POV to solve this. All animals, from single cells >>> to us, are fundamentally adaptive process control systems. We use our >>> intelligence to live better and more reliably, procreate, etc., much as >>> single-celled animals, only with MUCH richer functionality. Everything fits >>> this hierarchy of function leading to intelligence. >>> >>> Then, people like those on this forum start by ignoring this and trying to >>> create intelligence from whole cloth. This may be possible, but there is NO >>> existence proof for this, no data to guide the effort, etc. In short, there >>> is NO reason to expect a whole-cloth approach to work anytime during the >>> next century (or two). >>> >>> However, some of the mathematics of adaptive process control is known, and >>> I suspect the rest wouldn't be all that tough - if only SOMEONE were >>> working on it. > > > Erm.... guys. This would be me. > > I am working on it. For well over a decade now. Cognition and intelligence is > implemented as an adaptive control system replicating, inorganically, the > natural original called the human (mammal) nervous system. I simply replicate > it inorganically. Tough job but I am getting there. There's no programming. > No software. Just radically adaptively nested looping processes. In control > strategy terms it is a non-stationary system (architecture itself is > adaptive). Control loops come into existence and bifurcate and vanish > adaptively. The architecture commences at the level of single ion channels > and nest at multiple levels that then appear in tissue as neurons doing what > they do, but need not appear like this in the inorganic version. You don't > actually need cells at all. These then nest at increasing spatiotemporal > scales forming coalitions, layers, columns and finally whole tissue. All > inorganically. All the same at all scales from an adaptive control > perspective. Power-law scalable. Physically and logically. > > In my case, for the conscious version the hardware includes the > field-superposing, active additional feedback in the wave mechanics of the EM > field system produced by brain cells at specific points. The fields form an > addition/secondary loop modulation that operates orthogonally, > outside/through the space occupied by the chip substrate. > > What I am starting with is the 'zombie' or symbolically ungrounded version. > It doesn't produce the active field system (missing a whole control system > feedback mechanism) and uses supervised learning (externalised by a conscious > human trainer) to compensate for the loss of the natural role consciousness > has as an endogenous supervisor. It will, in the zombie form, underperform in > precisely the way all computer AGI underperforms. This is what is missing > when you use computers to do it all. You end up with a recipe (software) for > pulling Pinocchio's strings. Whereas my system bypasses the puppetry > altogether. It makes the little boy, not the puppet. > > However you view it, there's nothing else there in a brain except nested > loops that have power-law responses in two orthogonal axes: sensory and > cognitive. Adding the field system to the sensory axis (e.g. visual > experience) or part of the cognitive axis (e.g. emotional experience) provide > the active role for consciousness implemented through the causal impact of > the Lorentz force within the hardware. I suppose it'd be an 'adaptive control > loop' philosophy for cognition and 'EM field theory of consciousness' > combined. No computing needed whatever. Just like the brain. Most of the last > ten years has been spent figuring out the EM field bits! That I am now > omitting, knowing what I lose when I do that (i.e. consciousness). > > Teeny weeny Zombie version 0.0 this year I hope. No EM field generation. I > call it the 'circular causality controller'. I aim to add the EM fields > later. That part requires $millions. It's chip-foundry stuff. > > So chalk me in under this 'adaptive control loop' category for AGI > implementation please. I know this forum is a 'using computers to do AGI' > forum so I'll just continue to zip it. I haven't mentioned it much over the > years because it seems that most of you aren't interested in my approach. For > reference and for the record.... I am the 'AGI as adaptive control' guy. > > cheers > colin > >>> >>> >>> I suspect that when the answers are known, it will be a bit like spread >>> spectrum communications, where there is a payoff for complexity, but where >>> ultimately there is a substitute for designed-in complexity, e.g. like the >>> pseudo-random operation of spread spectrum systems. Genetics seems to >>> prefer designed-in complexity (like our brains) but there is NO need for >>> computers to have such limitations. >>> >>> Whatever path you take, you must "see a path" to have ANY chance of >>> succeeding. You must have a POV that helps you to "cut the crap" in pursuit >>> of your goal. Others here are working on whole-cloth approaches, yet >>> bristle when challenged for lacking a guiding POV. I see some hope in >>> adaptive control math. Perhaps you see something else, but it MUST have an >>> associated guiding POV for you to have any hope of succeeding - more than a >>> simple list of what it does NOT have. >>> >>> Steve ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/11721311-f886df0a Modify Your Subscription: https://www.listbox.com/member/?& Powered by Listbox: http://www.listbox.com AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
