Note to ZS readers: save this message for Monday, I just had to get it off my HD tonight. (too lazy to hold it).







om

Today I've decided, foolishly, to spill the beans on some of my more advanced concepts for AI. I'm going to cover the parts of my Hypermind architecture that I can remember right now and discuss their feasibility and the technical problems in implementing them on both hardware and as a bio-retrofit.

The basic capabilities I'll be discussing are distinct but closely related:

-> Algorithmic learning.
-> multi-session intelligence with interleaved sleep and wake sessions.
-> multi-configuration intelligence
-> distributed intelligence
-> High performance intelligence.

om

One of the annoyingist things about learning is repetition, or having to go through ginormous training sets to learn a simple idea. Neural nets, both natural and artificial use what is called adaptive learning. Obviously, it isn't all that great, even in the high performance neurons we humans get to use. What is going on is that our populations of neurons have to be slowly nudged into the shape we want them to be in. This takes practice, repetition, and lots of Dr. Pepper. (not that I use that specific drink).

What if we could learn algorithmically? What if we only needed to be exposed to enough input to reasonably disambiguate the information to be learned and then be able to use that knowledge, with masterful skill until we run out of disk space and are forced to prune back the less-useful stuff... The human memory system appears to be able to buffer information in short term memory long enough for the long-term neural nets to learn it, intellectually at least (with pathetically limited bandwidth)... Skill memory is much trickier because it involves wiring up what amount to new processor elements. So a student of math, at any level, must work through many related problems before the skill pathways begin to form, the long term memory is mostly useless for this kind of skill... The amount of practice required to master a musical instrument is legendary.

The first step is to identify what exactly needs to be learned. That is why the human brain relies on its imagination as it's primary mode of perception. It compares the actual scene with the imagined scene to identify surprising stimuli which it has special neurons rigged to detect and then focus as learning targets. The most annoying kinds of sounds are sounds that can't be modeled by the brain and therefore are constantly hitting this surprise circuit. The brain is certainly much more advanced than current neural networks but can we go any further? We need is a much more efficient means of generating motor programs. One of the issues is that the brain is pattern based, not math based, a math co-processor, much more precise and reliable than the cerebellum could provide the means for much more precise mental rehearsals. Furthermore, if the information is stored in discreet data structures rather than a matrix-blob as current neural networks are (or as a physically fixed cortical circuit), then an algebra and optimization process can be performed on them to much more quickly derive a more optimal representation that would otherwise take years to develop through practice.

Another problem with the matrix representation is that it simulates neural plasticity with a vector of weights. In larger networks, this vector is almost sure to be relatively sparse and therefore considerable amounts of computation will be wasted computing N*0 . Therefore the conventional systems are inherently unscaleable in addition to having the previously mentioned weaknesses.

In a biological context we need pattern processing elements that can be re-configured programatically and we need an organ that can inspect and re-organize these pattern processing elements on a continual basis.

om

One of the crippling limitations of our current neural architecture is that we can only do one thing at a time. The best we can do is rapidly time-share our brains between two tasks. This will almost inevitably lead to confusion as I frequently try to read an article and listen to a video talk or internet radio talk show at the same time. Our brains also must spend about 1/3rd of their time doing house-keeping tasks and simulations to rehearse behaviors and consolidate memories. What if we could treat our knowledge and memories as a mainframe hosted database and run tens of thousands of fully isolated concurrent processes over it. It might cut out useful forms of confusion but it would also allow us to fire up as many sessions as we need to process our 21st century information diet. This, again, requires us to be able to process memories and skills as discreet units that can be treated symbolically so that cached copies in different partitions can be re-integrated, as it has been noted that this won't work in neural nets.

So we need a cognitive architecture that can be multi-tasked out to the limit of the hardware... Conceptually, this is no different than how an operating system carries out multitasking, it is only necessary to be able to virtualize the hardware and run the virtual instances concurrently. Then RDBMS techniques can be applied to the stored abstractions and there you go...

In a biological context, it is necessary to be able to tag and index activation patterns so that they can be processed independently against the set of abstractions.

om

When one thinks about all the kinds of things one would want to extend one's mind into in all of transhumandom (real world applications always being more interesting than VR...) one is boggled by the variety. On the extreme, you might want a highly optimized flame-bot for getting into IRC and mailing list flame-wars, its modalities would be primarily text-based and would only have enough visualization to be able to properly interpret the language. On the other extreme you would might have a very large and highly complex virtual body with many conventional and unconventional modalities for 3rd generation cyber-sex... Both systems would act concurrently on the same database... The basic problem is that you need to select and configure an ad-hoc network of functional units as needed for the selected embodyment. The fixed white-matter pathways one finds in conventional brains (and simulations thereof) hardly provides for any flexibility. The only solution is to chuck it and use computer data structures.

A biologically embodied superintelligence wouldn't need to deal with as much change as mentioned above. It would be sufficient that it merely have a few spare general-purpose channels that can be tasked as necessary. Additional capabilities would come through networking in semi-sentient sub-units.

om

That brings us to distributed intelligence. For many reasons it is desirable to spread a mind across geographically separated locations. The physical security of any one site can never be perfect and it is probable that low-latency interactions will be desirable in a number of contexts. The first issue is latency, the second issue is the one Eugene likes to beat us over the head with, synchronizing slower hosts with faster hosts. There are two basic types of interactions here. The first is the knowledge, skills, and memories. This level of knowledge, as I proposed above, can be treated like a database and therefore the standard synchronization techniques can be applied. Latency is not a major problem though bandwidth is. As bandwidth becomes choked, each side the link will have to prioritize traffic. This is not a problem at all. What is a problem is the high level executive function, the thread of consciousness who's modality is to manage all of the other parallel operations. As the system becomes larger, it will be necessary to apply a deeper and deeper hierarchy of executive functions. Unless there's a major breakthrough in quantum communication, it appears that it will be necessary that the top level executive function will necessarily be quite slow. However, this does not place any constraints on the subordinate exeucitives, or the workers. Unlike what you may have read in sci-fi, the parts of the AI you actually interact with locally will run at full speed but they will not have any special privileges in communicating with other agents within the same mind on a distant planet.

Eugene says that neural interfaces are pointless because human neurology is so many orders of magnitude slower than an AI substrate (actually, he claims this is true of uploads too!!!). He never goes past the quick put-down. He assumes what he perceives to be perfectly obvious to be an insurmountable, unavoidable obstacle. The question is, exactly how much communication do you need between the "fast" parts of your mind and the legacy parts of your mind? Is there sufficient bandwidth available and does the latency create a bottleneck?

As I explained above, when bandwidth is limited, then only the most important messages are sent. This is an inherent limitation in the human brain that simply cannot be overcome by any conceivable technique, including uploading, because the neural architecture is too limited in capacity and has the previously mentioned limitations in learning and scalability. It is also acknowledged that a slow-time or even uploaded human would have difficulty managing more than a small number of interlinked agents. The only way around that is to use a more powerful executive function external to the base neural architecture. This does have the side effect of reducing you to an appendage of a larger being but I'm pretty sure that's the best deal you can possibly get. The best part of it, you get to live through the process!

om

Now for the last part of the discussion, pushing performance through the roof. "clock speed" and even basic algorithmic tweaks are too trivial to mention. Where the real gold is, is in expanding the architecture to be able to grok a chemistry database of 2 * 10^6 unique chemicals. Or truly understand 5-space physics. What you need is the ability to manipulate and extend the architecture of the mind as the cognitive demands arise. It's difficult to say when a new domain sub-processor would be required... Maybe a frustration detector would be able to identify where the existing system is failing, or maybe the problem would be optimized out by assuming everything is higher-dimensional from the outset and only reducing dimensions as the true eigen-spaces become apparent. Regardless, the true power of superintelligence is the ability to expand and reconfigure as the need arises. This approach, and only this approach is capable of tackling the grand challenges of the universe.



Conclusion: superintelligence looks nothing like an upload! =P



--
E T F
N H E
D E D

Powers are not rights.



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968
Powered by Listbox: http://www.listbox.com

Reply via email to