Prednet develops consciousness?

On Wed, Jun 19, 2019, 06:51 Alan Grimes via AGI <agi@agi.topicbox.com>
wrote:

> Yay, it seems peeps are finally ready to talk about this!! =P
> 
> Lets see if I can fool anyone into thinking I'm actually making sense by
> starting with a first principles approach...
> 
> On first approximation, yer computer's memory is just a giant string of
> charactors. Since everything you have ever done on your computer has
> been encoded into this string, we can be reasonably confident that this
> is a universal way to encode anything, so we go ahead and say that all
> the modalities (sight, sound, etc,) are mapped to this memory and
> continue on.
> 
> [I'm maximally overtired right now, I didn't really get any sleep last
> night, should wait until tomorrow at least to write this but I want to
> get a head start...]
> 
> If you prefer, you can think in terms of a multi-tape turing machine,
> where the program and working variables are stored on other tapes and
> the mind file is in main memory, but the point is that we presume there
> exists a computation that does intelligence using the main memory, which
> we will say is the complete state of the mind.
> 
> So therefore what we are looking for is a set of principles and basic
> operations that, when applied to the main memory, intelligent behavior
> and subjective experience emerges.
> 
> [next evening]
> 
> The most important thing to realize is that the data is NOT
> RANDOM/stochastic/whatever, it is STRUCTURED,?? It has patterns, it has
> types, it may have meaning.?? This is true of all modalities.
> 
> Therefore, in order to create a mind, we need the system to be able to
> capture that structure and be able to use it in computations. Lets say
> that we had some identified structure in the data which had some modest
> but non-trivial k-complexity. Replacing that data with the K-complexity
> description, even if if were a "good enough" approximation, would
> satisify most definitions of understanding that structure and creating
> some way to point to that description and incorporate it in other
> descriptions would satisfy the definition of abstraction.
> 
> I hope you are all boned up on your study of non-standard computational
> models. What I'm getting at is that the expression of the K-complexity
> is a structured object just like the structured object which it expands
> to. So we can treat it the same way and approximate it using the basic
> functions we use to manipulate abstractions.
> 
> So if you had a string/object that basically meant "make big" and
> another that expressed the shape, then you only need a way to apply one
> to the other to make a big shape. It is important that things are done
> like this because it is how general intelligence is achieved, because it
> allows the operation as well as the object to be learned by essentially
> the same mechanism.
> 
> Now, let me shift gears here and go into a thread that I'd been meaning
> to start for the past week or so. DEEP LEARNING IS DEAD! If we don't
> have a viable replacement for it in the next 18 months or so, we're at
> risk of another AI winter. =0???? There are certainly some good and useful
> machine learning algorithms out there, some of which are just
> rediscovery of techniques from ages past, but well that's OK too.
> 
> When you get down to it, DL is the realization of ideas dating back a
> very long time but only made feasible with gpGPU techniques, see
> Nvidia's website for a recounting of the history.
> 
> [few days later]
> 
> Have a look at this and count up the number of problems this field of
> study has from an AGI perspective:
> 
> https://scholar.google.com/scholar?q=neural+architecture+search&hl=en&as_sdt=0&as_vis=1&oi=scholart
> 
> Neural nets are brittile in that you can only select their topology when
> you are creating them. (iirc)
> 
> The brain, by contrast is quite infamous for exhibiting plasticity in
> the face of dammage and/or changing requirements. For example, london
> taxi drivers develop an enlarged hypocampus to help deal with the
> maze-like roads in that city.
> 
> https://www.youtube.com/watch?v=UoJf_tXU2Zk
> 
> Ok, that video was ammusing and little else. ;)
> 
> Also, consider what a "golden network" would look like. A golden network
> is a network like a universal turing machine that exhibits general AI
> instead of some special AI.
> 
> The bottom line is that while we have come a long way from 2012, we are
> reaching the top of the S-curve with respect to deep learning. Once the
> RoI falls below some critical level, we will enter an AI winter. =|
> 
> We do not have time for another winter.
> 
> Deep learning is also hitting a computational complexity wall but I
> couldn't find good referances for that so I left it out of this post.
> 
> Here's something else to think about.
> 
> Consider the near photorealistic graphics that you got with Assassin's
> Creed on the Playstation 3.
> 
> Consider the amount of power people are throwing at autonomous driving
> and just recognising the general outlines of visual scenes.
> 
> One of the reasons I flog prednet so much,
> 
> https://coxlab.github.io/prednet/
> 
> is not just because it yields consciousness, but it reduces perception
> to something a PS3 could do.
> 
> You render what you think you are seeing, the LGN of the thalamus
> computes the error signal, and then you use your expensive complex
> computations or just that error signal!
> 
> [midnight a day or two later, just jamming random related topics into
> the post to make myself look smart...]
> 
> Here's another fascinating area of study:
> 
> https://scholar.google.com/scholar?q=explainable+neural+networks&hl=en&as_sdt=0&as_vis=1&oi=scholart
> 
> To me, it's just proof that deep learning, despite what can be
> accomplished with it, is NOT a form of intelligence. When you think
> about it, it doesn't learn in any true sense, in that it can recite
> facts, or recall events but it, over many thousands of iterations, can
> *adapt*.
> 
> The result is a system that essentially operates as mindlessly as a
> reflex. The apparent complexity of the behavior not changing that fact.
> 
> Going to a recurrent model does give the network a legit memory, after a
> fassion. Still, the prospects for further progress is dim.
> 
> [next night around midnight again]
> 
> Consider, by contrast a system that works by analyzing structure. There
> is no reason why it can't simply continuallly fill its memory, because
> each thing is recorded as a representation of it's structure, the
> storage requirement for even millenia of ordinary life experience is not
> at all implausible.
> 
> Ok, the magic sauce here is an ability to organize and catalog
> representations. I think there exists a geometric technique that can
> yield pattern recognition. This would involve a way to normalize the
> input and then search the database, the transformation used to achieve
> the normalization also being a relevant structure that is also analyzed
> in the same manner.
> 
> Other tricks such as the space-time rotation that I've described a
> number of times are also important techniques for detecting spatial
> patterns. I think that approach can be much more powerful than
> "convolution" which is just scanning a larger matrix with a smaller...
> I'm talking about taking the visual scene and turning the scan of it
> into a temporal signal like a sound. Just as a speech recognizer doesn't
> care when the speaking actually begins, treating spatial signals can
> work the same way.
> 
> Anyway, what was this post supposed to be about again? bleh -> hitting
> send.
> 
> --
> Please report bounces from this address to...@numentics.com
> 
> Powers are not rights.
> 

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/T395236743964cb4b-M686d9fcf7662ad8dc2fc1130
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to