Re: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-25 Thread Brad Wyble
 
 This is exactly backward, and which makes using it as an unqualified
 presumption a little odd.  Fetching an object from true RAM is substantially
 more expensive than executing an instruction in the CPU, and the gap has
 only gotten worse with time.


That wasn't my point, which you may have missed.  The point is that with
our current technology track it's far cheaper to double your memory
than to double your CPU speed.  I'm not referring to the amount of memory
bits processed by the CPU, but the total number of pigeonholes available.  
These are not one and the same.

Therefore you can make gains in representational power by boosting the
amount of RAM, and having each bit of memory be a more precise
representation.  You can afford to have, for example, a neuron encoding
blue sofas and a neuron encoding red sofas.  While a more restricted RAM
approach would need to rely on a distributed representation, one with only
sofa neurons and color neurons. (apologies for the poor example, but I'm 
in a hurry)

Your points are correct, but refer to the bottleneck of getting 
information from RAM to the CPU, not on the total amount of RAM available.  


 Back to the problem of the human brain, a big part of the problem in the
 silicon case is that the memory is too far from the processing which adds
 hard latency to the system.  The human brain has the opposite problem, the
 processing is done in the same place as the memory it operates on (great
 for latency), but the operational speed of the processing architecture is
 fundamentally very slow.  The reason the brain seems so fast compared to
 silicon for many tasks is that the brain can support a spectacular number of
 effective memory accesses per second that silicon can't touch.

Both technologies have their advantages and disadvantages.  The brain's
memory capacity (in terms of number of addressable bits) cannot be
increased easily while a computer's can be.  I merely suggest that this
fundamental difference is something to consider if one is intent on
implementing AGI in a Neumann architechture.




---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


Re: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-24 Thread deering



Ben, you haven't given us an update on how things 
are going with the Novamente A.I. engine lately. Is this because progress 
has been slow and there is nothing much to report, or you don't want to get 
peoples hopes up while you are still so far from being done, or that you want to 
surprise us one day with, "Hey guys, guess what? The Singularity has 
arrived!"





To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]




RE: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-24 Thread Ben Goertzel


Brad,

Hmmm... yeah, the problem you describe is actually an implementation issue,
which is irrelevant to whether one does synchronoous or asynchronous
updating.

It's easy to use a software design where, when a neuron sends activation to
another neuron, a check is done as to whether the target neuron is over
threshold.  If it's over threshold, then it's put on the ready to fire
queue.  Rather than iterating through all neurons in each cycle, one simply
iterates through those neurons on the ready-to-fire queue.

Of course, one can use this approach with either synchronous or asynchronous
updating.

We used this design pattern in Webmind, which had a neural net aspect to its
design; Novamente is a bit different, so such a strategy isn't relevant.

-- Ben G


 While I haven't read any of the documents in question, I'd like
 to expound
 a bit here.

 While you are certainly correct, I think Pei was referring to the wasted
 computational power of updating synapses that are inactive and have no
 chance of being activated in the near future.  In our current Neumann
 architectures, memory is much cheaper than CPU cycles, which is
 not the case in the brain.

 So while the brain opts for minimal neurons, and keeps most of them active
 in any given situation, a silicon NN might have factors of 10 more
 neurons, but use very sparse encoding and a well optimized update
 algorithm.  This setup would emphasize only spending CPU time updating
 neurons that have a chance of being active.


 -Brad

 ---
 To unsubscribe, change your address, or temporarily deactivate
 your subscription,
 please go to http://v2.listbox.com/member/[EMAIL PROTECTED]




---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


RE: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-24 Thread Brad Wyble


Guess I'm too used to more biophysical models in which that approach won't
work.  In the models I've used (which I understand aren't relevant to your
approach) you can't afford to ignore a neuron or its synapses because they
are under threshold.  Interesting dynamics are occurring even when the
neuron isn't firing.  You could ignore some neurons that are at rest and 
hadn't received any direct or modulatory input for some time, but
then you'd need some fancy optimizations to ensure you're not missing
anything.

But in the situation you're referring to with a more abstract (and 
therefore more useful to AGI) implementation, these details are 
irrelevant.  

I just wanted to chime in and ramble a bit :)

Very glad to hear things are going well with Novamente.  

Hope the holidays treat all of you well.

-Brad



---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


RE: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-24 Thread Ben Goertzel

Yep, you're right of course.  The trick I described is workable only for
simplified formal NN models, and for formal-NN-like systems such as Webmind.
It doesn't work for neural nets that more closely simulate physiology, and
it also isn't relevant to systems like Novamente that are less NN-like

ben

 -Original Message-
 From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
 Behalf Of Brad Wyble
 Sent: Wednesday, December 24, 2003 11:59 AM
 To: [EMAIL PROTECTED]
 Subject: RE: [agi] The emergence of probabilistic inference from hebbian
 learning in neural nets




 Guess I'm too used to more biophysical models in which that approach won't
 work.  In the models I've used (which I understand aren't relevant to your
 approach) you can't afford to ignore a neuron or its synapses because they
 are under threshold.  Interesting dynamics are occurring even when the
 neuron isn't firing.  You could ignore some neurons that are at rest and
 hadn't received any direct or modulatory input for some time, but
 then you'd need some fancy optimizations to ensure you're not missing
 anything.

 But in the situation you're referring to with a more abstract (and
 therefore more useful to AGI) implementation, these details are
 irrelevant.

 I just wanted to chime in and ramble a bit :)

 Very glad to hear things are going well with Novamente.

 Hope the holidays treat all of you well.

 -Brad



 ---
 To unsubscribe, change your address, or temporarily deactivate
 your subscription,
 please go to http://v2.listbox.com/member/[EMAIL PROTECTED]




---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


Re: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-23 Thread Pei Wang
 Actually, in attractor neural nets it's well-known that using random
 asynchronous updating instead of deterministic synchronous updating does
NOT
 change the dynamics of a neural network significantly.  The attractors are
 the same and the path of approach to an attractor is about the same.  The
 order of updating turns out not to be a big deal in ANN's.  It may be a
 bigger deal in backprop neural nets and the like, but those sorts of
neural
 nets are a lot further from anything I'm interested in...

I'd rather get ride of the notion of attractor altogether. Though it may
be useful for perception, in high-level cognition I don't see anything like
it. Of course, some beliefs are more stable than others, but are they states
to which all processes converge?

 Hmmm  Pei, I don't see how to get NARS' truth value functions out of
an
 underlying neural network model.  I'd love to see the details  If
truth
 value is not related to frequency nor to synaptic conductance, then how is
 it reflected in the NN?

What I mean is not that NARS, as a reasoning system, can be (partially or
completely) implemented by a network, but that NARS can be seen as a
network --- though different from conventional NN.

I think NN is much better than traditional AI in its philosophy --- I like
parallel processing, distributed representation (to a certain extent),
incremental learning, competing results, and so on. However, ironically, the
techniques of NN is less flexible than symbolic AI. I don't like NN when it
uses fixed network topology, has no semantics (and even claims it to be an
advantage), takes the goal of learning as converging to a function (mapping
input to output), does global updating, uses activation for both logical
and control purposes, and so on.

My way to combine the two paradigms is not to build a hybrid system that is
part symbolic and part connectionist, but to build a unified system which is
similar to symbolic AI in certain aspects, and similar to NN in some other
aspects.

Pei


---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


RE: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-23 Thread Ben Goertzel

Hi,

 Actually, in attractor neural nets it's well-known that using random
  asynchronous updating instead of deterministic synchronous updating does
 NOT
  change the dynamics of a neural network significantly.  The
 attractors are
  the same and the path of approach to an attractor is about the
 same.  The
  order of updating turns out not to be a big deal in ANN's.  It may be a
  bigger deal in backprop neural nets and the like, but those sorts of
 neural
  nets are a lot further from anything I'm interested in...

 I'd rather get ride of the notion of attractor altogether. Though it may
 be useful for perception, in high-level cognition I don't see
 anything like
 it. Of course, some beliefs are more stable than others, but are
 they states
 to which all processes converge?

You have a point, which si why I didn't use the term attractor in the
Hebbian Logic paper.

The results I cited about attractor neural nets have to do with attractors.

But in the brain, or in a Hebbian Logic network, you don't have
attractors -- what you have are probabilistically invariant subsets of
state space, i.e. subsets of the system's state space with the property
that, once a system gets in there, it's likely to stay there a while.
Attractors are a limiting case of this kind of state-space-subset, and
they're a limiting case that doesn't occur in the cognitive domain.

  Hmmm  Pei, I don't see how to get NARS' truth value functions out of
 an
  underlying neural network model.  I'd love to see the details  If
 truth
  value is not related to frequency nor to synaptic conductance,
 then how is
  it reflected in the NN?

 What I mean is not that NARS, as a reasoning system, can be (partially or
 completely) implemented by a network, but that NARS can be seen as a
 network --- though different from conventional NN.

 I think NN is much better than traditional AI in its philosophy --- I like
 parallel processing, distributed representation (to a certain extent),
 incremental learning, competing results, and so on. However,
 ironically, the
 techniques of NN is less flexible than symbolic AI. I don't like
 NN when it
 uses fixed network topology, has no semantics (and even claims it to be an
 advantage), takes the goal of learning as converging to a
 function (mapping
 input to output), does global updating, uses activation for both logical
 and control purposes, and so on.

 My way to combine the two paradigms is not to build a hybrid
 system that is
 part symbolic and part connectionist, but to build a unified
 system which is
 similar to symbolic AI in certain aspects, and similar to NN in some other
 aspects.

Firstly, Novamente is not a hybrid system that's part symbolic and part
connectionist, either.

Webmind was, but Novamente isn't anymore.  There's no more association
spreading or activation spreading; these NN-like processes have been
replaced by specialized deployments of PTL (probabilistic reasoning
Novamente-style).

Novamente does hybridize a bunch of things: BOA learning, combinatory logic,
PTL inference, etc. ... but not any NN stuff

Second, I do not advocate neural nets as an approach to AI, either.  I think
the approach has its merits, but overall I think that NN's are a really
inefficient way to use von Neumann hardware.

If we knew enough to *really* emulate the brain's NN in software, then the
guidance offered by the brain would be so valuable as to offset the
inefficiency of implementing massively-parallel-wetware-oriented structures
and algorithms on von Neumann hardware.  But we don't know nearly enough
about the brain to make brain-emulating NN's; and the currently popular NN
architectures seem to satisfy neither the goal of brain emulation, nor the
goal of efficient/effective AI.

My point in articulating Hebbian Logic is NOT to propose it as an optimally
effective approach to AI, but rather to propose it as a conceptual solution
to the conceptual problem of: **How the hell do logical inference and
related stuff emerge from neural networks and other brainlike stuff?**

No one in cognitive science seems to have a good explanation of this, beyond
the really vague handwaving level.  I think that the Hebbian Logic approach
provides a significantly better explanation than anyone else has given so
far.  Even given that it also involves a bunch of handwaving (since I didn't
work out all the technical details, and probably won't do so soon due to my
own time limitations).

Hebbian Logic *might* be a decent approach to practical AI --- I don't think
it would be a terribly stupid approach --- but I like the Novamente approach
better...

-- Ben


---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


Re: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-22 Thread Pei Wang
Ben,

Some comments to this interesting article:

*. S = space of formal synapses, each one of which is identified with a
pair (x,y), with x Î N and y ÎNÈS.

Why not x ÎNÈS?

*. outgoing: N à S* and incoming: N - S*

Don't you want them to cover higher-order synapses?

*. standard neural net update and learning functions

One thing I don't like in NN is globle updating, that is, all activations
and weights are updated in every step. Even if it is biologically plausible
(which I'm not sure), in an AI system it won't scale up. I know to drop this
will completely change the dynamics of NN.

*. AàI B, means that when B is present, A is also present

What are A and B (outside the network)? Are they terms, sets, attributes,
events, or propositions? What do you mean by present?

*. probability P(A,t), defined as the probability that, at time t, a
randomly chosen neuron xÎA is firing and the conditional probability
P(A|B; t) = P(A ÇB,t)/ P(B,t)

This is the key assumption made in your approach: to take the frequency of
firing as the degree of truth. I need to explore further about its
implications, though currently I feel uncomfortable. In my own network
interpretation of NARS (for a brief description, see
http://www.cogsci.indiana.edu/farg/peiwang/papers.html#thesis Section 7.5),
I take activation/firing as a control parameter, indicate the recourse
spends on the node, which is independent to the truth value --- I'm
thinking about T and T is true are fundamentally different.

Of course, the logic/control distinction is not in NN, where both are more
or less reflected in activation value. When you map their notions into
logic, such a distinction become tricky.

*. Basic inference rules

I don't see what is gained by a network implementation (compared to direct
probabilistic calculation).

*. Hebbian Learning

The original Hebbian learning rule woks on symmetric links (similarity, not
inheritance), because weight of a link is decrease when one end is activated
and the other isn't, and which is which doesn't matter.  What you does in
Hebbian learning variant A is necessary, but it is not the original
Hebbian learning rule.

*. Section 6

I'm not sure I understand the big picture here. Which of the following is
correct?

(1) PTL is fully justified according to probability theory, and the NN
mechanism is used to implement the truth value functions.

(2) PTL is fully justified according to probability theory, and the truth
value functions are directly calculated, but the NN mechanism is used to
implement inference control, that is, the selection of rules and premises in
each step.

(3) The logic is partially justified/calculated according to probability
theory, and partially according to NN (such as the Hebbian learning rule).

*. In general, I agree that it is possible to unify Hebbian network with
multi-valued term logic (with an experience-grounded semantics). NARS is
exactly such a logic, where a statement is a link from one term to another,
and its truth value is the accumulated confirmation/disconfirmation record
about the relation. In NARS, Hebbian learning rule correspond to the
comparison (with induction, abduction, and deduction as variants) plus
revision. Activation spreading corresponds to (time) resource allocation.

BTW, Pavlov's conditioning is similar to Hebbian learning, and can also be
seen as special case of induction in (higher-order) multi-valued term logic.

Pei

- Original Message - 
From: Ben Goertzel [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Saturday, December 20, 2003 8:26 PM
Subject: [agi] The emergence of probabilistic inference from hebbian
learning in neural nets



 Hi,

 For those with the combination of technical knowledge and patience
required
 to sift through some fairly mathematical and moderately speculative
cog-sci
 arguments... some recent thoughts of mine have been posted at

 http://www.goertzel.org/dynapsyc/2003/HebbianLogic03.htm

 The topic is:
 **How to construct a neural network so that symbolic logical inference
will
 emerge from its dynamics?**

 This is not directly relevant to my own current AI work (Novamente,
 www.agiri.org), which is not neural network based.  However, it is
 conceptually related to Novamente; and more strongly conceptually related
to
 Webmind, the previous AGI design with which I was involved.  It is also
 loosely related to Pei Wang's NARS inference system.

 While my guess is that this is not the most effective path to AGI at
 present, I do think it's a very interesting area for research and an
 exploration-worthy potential path toward AGI.

 Apologies for the rough-draft-ish-in-places document formatting ;-)

 -- Ben


 ---
 To unsubscribe, change your address, or temporarily deactivate your
subscription,
 please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


---
To unsubscribe, change your address, or temporarily deactivate your subscription, 
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]


RE: [agi] The emergence of probabilistic inference from hebbian learning in neural nets

2003-12-22 Thread Ben Goertzel

Pei,

Thanks for your thoughtful comments!  Here are some responses...

-
*. S = space of formal synapses, each one of which is identified with a
pair (x,y), with x Î N and y ÎNÈS.

Why not x ÎNÈS?
-

No strong reason -- but, I couldn't see a need for that degree of generality
in a Hebbian Logic context...
of course, there's no reason not to allow it in a formal model.

---
*. outgoing: N à S* and incoming: N - S*

Don't you want them to cover higher-order synapses?
---

Yeah, you're right.

However, I may remove higher-order synapses from the paper entirely,
preferring to deal with higher-order relations via links to multi-neuron
paths as discussed later on.


-
*. standard neural net update and learning functions

One thing I don't like in NN is globle updating, that is, all activations
and weights are updated in every step. Even if it is biologically plausible
(which I'm not sure), in an AI system it won't scale up. I know to drop this
will completely change the dynamics of NN.
--

Actually, in attractor neural nets it's well-known that using random
asynchronous updating instead of deterministic synchronous updating does NOT
change the dynamics of a neural network significantly.  The attractors are
the same and the path of approach to an attractor is about the same.  The
order of updating turns out not to be a big deal in ANN's.  It may be a
bigger deal in backprop neural nets and the like, but those sorts of neural
nets are a lot further from anything I'm interested in...

---
*. probability P(A,t), defined as the probability that, at time t, a
randomly chosen neuron xÎA is firing and the conditional probability
P(A|B; t) = P(A ÇB,t)/ P(B,t)

This is the key assumption made in your approach: to take the frequency of
firing as the degree of truth. I need to explore further about its
implications, though currently I feel uncomfortable. In my own network
interpretation of NARS (for a brief description, see
http://www.cogsci.indiana.edu/farg/peiwang/papers.html#thesis Section 7.5),
I take activation/firing as a control parameter, indicate the recourse
spends on the node, which is independent to the truth value --- I'm
thinking about T and T is true are fundamentally different.

Of course, the logic/control distinction is not in NN, where both are more
or less reflected in activation value. When you map their notions into
logic, such a distinction become tricky.
---

Yeah, to make Hebbian Logic work, you need to assume that frequency of
firing roughly corresponds to degree of truth -- at least, for those neural
clusters that directly represent symbolic information.

So, for instance, the cat cluster fires a lot when a real or imaginary cat
is present to the mind.

If the mind wants to allocate attention to the cat cluster, but there is
no real cat present, it must then either

-- find a way to stimulate other things logically related to cat
-- create abstract quasi-perceptual stimuli that constitute a mock cat and
fool the cat cluster into firing

I think this is how the brain and human mind work.  I agree it's not
optimal, and that in an AI system it's nicer to make separate parameters for
activation and truth value, as is done in both NARS and Novamente.

-
*. Basic inference rules

I don't see what is gained by a network implementation (compared to direct
probabilistic calculation).
-

Actually, I think there is no big advantage.  This issue is discussed in the
very last section of the paper.  My view is that the brain uses a horribly
inefficient mechanism to achieve probabilistic inference, and AI systems can
achieve the same thing more efficiently.

I prefer the Novamente implementation of PTL to a Hebbian Logic
implementation.  However, I think it's interesting to observe,
theoretically, that a Hebbian Logic representation is possible.

-
*. Hebbian Learning

The original Hebbian learning rule woks on symmetric links (similarity, not
inheritance), because weight of a link is decrease when one end is activated
and the other isn't, and which is which doesn't matter.  What you does in
Hebbian learning variant A is necessary, but it is not the original
Hebbian learning rule.


Oops, you are right.  My variant A is fairly standard in the literature
these days, but it's not the original one.  I will correct that, thanks.


*. Section 6

I'm not sure I understand the big picture here. Which of the following is
correct?

(1) PTL is fully justified according to probability theory, and the NN
mechanism is used to implement the truth value functions.

(2) PTL is fully justified according to probability theory, and the truth
value functions are directly calculated, but the NN mechanism is used to
implement inference control, that is, the selection of rules and premises in
each step.

(3) The logic is partially justified/calculated according to probability
theory, and partially according to