Sergio -

These two links should get you started. They contain many resources
explaining the CLA and HTM:

https://github.com/numenta/nupic/wiki/CLA-and-HTM-Theory
http://numenta.com/learn/

Regards,

Jeff Fohl

On Sat, Jan 10, 2015 at 7:24 AM, sergio marinho <[email protected]>
wrote:

> Hi, fellows.
>
> Where can I find the last papers, the novelties  on HTM and CLA?
>
> Regards,
>
> Sergio Soares.
>
>
>
>
>
>   Em Sexta-feira, 9 de Janeiro de 2015 22:04, Subutai Ahmad <
> [email protected]> escreveu:
>
>
>
> Very nice description Fergal! It's worth emphasizing your last point. The
> term HTM is more meaningful and should be used outside the NuPIC community.
> We're trying to stop using the term "CLA" altogether.
>
> --Subutai
>
> On Fri, Jan 9, 2015 at 5:27 AM, Fergal Byrne <[email protected]>
> wrote:
>
> Hi Dinesh,
>
> HTM refers to the general theory developed by Jeff Hawkins and Numenta
> over the past 1-15 years. You can think of HTM as the general "big idea" of
> how we believe the neocortex works. The key aspects of HTM are Jeff's six
> principles, which refer to hierarchy, sparse distributed representations,
> online learning from streaming data, a uniform algorithm, combination of
> sensory and motor function everywhere, and attention. While the theory will
> accumulate detail (for example what the roles of the layers inside a region
> might be doing), it grows outwards stably from this kernel.
>
> Officially, CLA refers to the particular detailed algorithmic design for a
> single layer of neurons, which is outlined in the 2011 White Paper and
> (partially) implemented by NuPIC. Jeff Hawkins and Numenta have indicated
> that they wish to "freeze" this meaning of CLA and use a different name for
> new versions of their detailed algorithmic designs.
>
> The rest of us have become accustomed to using "CLA" to refer to an
> algorithmic design which is close to Numenta's, but might differ in some
> minor or major aspects. The key features of CLA, which generalise across
> most of our models, are:
>
> - Neurons arranged in columns ("mini-columns" in neocortex) which share
> feedforward inputs and have similar feedforward responses.
> - Sparsity imposed by a columnar inhibition algorithm.
> - Feedforward inputs appear on proximal dendrites (to a column in official
> CLA, also to cells in some models).
> - Neurons in a layer have axons connected to distal dendrites in the same
> layer, allowing for prediction.
> - Proximal dendrites perform some version of linear summing.
> - Distal dendrite segments act independently as coincidence detectors.
> - Layers can learn first-order transitions between feedforward patterns,
> and also higher-order sequences using choices of active cells in an active
> column.
> - Columns which correctly predict their activity have one cell active,
> otherwise several cells activate (burst).
>
> HTM is quite general, allowing for many more detailed theories and designs
> to be claimed to correspond to HTM, but It's much easier to quantify how
> well a design matches up with CLA proper.
>
> We tend to use CLA when referring to processes in some detail (at the
> layer, column, neuron, dendrite, synapse levels), and HTM when talking
> about how things work at the layer, region and brain levels. We'll also be
> seen using "HTM" when we propose ideas which supercede or contradict
> assumptions underlying Numenta's "official" CLA design.
>
> The other thing to bear in mind is that CLA is an internal name (within
> the community) which has no general currency in either neuroscience or
> AI/ML, while HTM is well-known (at least by name) to researchers in both
> fields.
>
> Regards,
>
> Fergal Byrne
>
>
> On Fri, Jan 9, 2015 at 12:48 PM, David Ragazzi <[email protected]>
> wrote:
>
> Dear Dinesh,
>
> > 1.What is the difference between CLA and HTM? 2.Is CLA generalization of
> HTM as the CLA(the agorithms based on cortex) name suggests so? Explain
> if wrong.
>
> CLA => Cortical Learning **ALGORITHMS**
> HTM => Hierarchical Temporal **THEORY**
>
> As the names say, CLA tries simulate what the HTM states about how cortex
> could work. Something we use wrongly HTM acronym to refer to CLA. But the
> names are clear, one is the theory, the other is the algorithmic model of
> it. Just remember neither all features addressed on HTM are implemented on
> CLA (yet).
>
> David
>
> On 9 January 2015 at 10:08, Dinesh Deshmukh <[email protected]> wrote:
>
> Hi
>
> 1.What is the difference between CLA and HTM?
> 2.Is CLA generalization of HTM as the CLA(the agorithms based on cortex) name
> suggests so?Explain if wrong.
>
> Thank you.
>
>
>
>
> --
> David Ragazzi
> MSc in Sofware Engineer (University of Liverpool)
> OS Community Commiter at Numenta.org
> --
> "I think James Connolly, the Irish revolutionary, is right when he says that
> the only prophets are those who make their future. So we're not
> anticipating, we're working for it."
>
>
>
>
> --
>
> Fergal Byrne, Brenter IT
>
> http://inbits.com - Better Living through Thoughtful Technology
> http://ie.linkedin.com/in/fergbyrne/ - https://github.com/fergalbyrne
>
> Founder of Clortex: HTM in Clojure -
> https://github.com/nupic-community/clortex
>
> Author, Real Machine Intelligence with Clortex and NuPIC
> Read for free or buy the book at https://leanpub.com/realsmartmachines
>
> Speaking on Clortex and HTM/CLA at euroClojure Krakow, June 2014:
> http://euroclojure.com/2014/
> and at LambdaJam Chicago, July 2014: http://www.lambdajam.com
>
> e:[email protected] t:+353 83 4214179
> Join the quest for Machine Intelligence at http://numenta.org
> Formerly of Adnet [email protected] http://www.adnet.ie
>
>
>
>
>

Reply via email to