Hi,

Yes, copycat simulated a metric in an ad hoc way, because it lacked a
robust way of measuring and utilizing uncertainty...

I am unsure (heh heh) what uncertainty has to do with it. CC got a fixed, completely known problem. It could only construct valid interpretations, so there was no doubt that any given one was correct. It had no memory for previous problems, so it had no principled way of deriving a prior for any given interpretation or technique (nor did it even have a separate concept of technique, except as a set of weights on its higher-level concepts such as "opposite").

Can you describe how you'd (in theory) enhance CC by incorporating measures of uncertainty?


Without taking a lot of time (maybe I'll elaborate more later), the point is that humans solve analogy problems not (usually) by finding specific strong analogies, but by finding a huge number of weak analogies and statistically polling the ensemble.... Even when there is a strong analogy, it is usually bolstered by a whole lot of weak analogies surrounding it in "possible analogy space"...

To be at all intelligence, CC needs a memory of MANY previous problems, so that it can make many weak probabilistic matches to prior problems, which can guide it through the large space of possible transformations of the current problem toward a small set of potentially appropriate ones

I.e., CC does not use long-term memory to probabilistically guide the "inference tree pruning" process, ergo it gets lost in combinatorial explosions and can't find the right matches or recognize (via statistical comparison to rightness in prior situations) their rightness...

The sense in which the "right" analogies are Occam programs is only statistical: they are "right" because the patterns they embody simplify a LOT of different problems.... Their rightness cannot be identified effectively by a system without long-term memory, and an effective LTM requires
effective probabilistic inference.

Anyway, the interesting thing in Hofstadter-ian analogy problems is the
identification of
WHAT THE SALIENT ATTRIBUTES ARE ... not the choice of representation
itself.

I must respectfully but implacably disagree. In (a hypothetical) Supercat, which CC dreamt of being when it grew up, the salient attributes are exactly the representation, which are invented by the system under the pressures of the particular problem. (CC itself was aimed this way but did perhaps more choosing than inventing.) DH's entire point (the FARGument, if you will) with the whole feline series was that you can't represent and then match -- finding the representation is the hard part, and matching is trivial thereafter.

Sure ... I think we are just using language in different ways, and it may be my usage that's
nonstandard.

Hofstadter wrote about knob-creaton and knob-twiddling. I guess we both agree with him that the hard part is figuring out what knobs need to be twiddled and creating them.


If you represent concepts as probabilistic-logical combinations of the
salient attributes, you
will get the same advantages as in your numerical representation plus
more ;-)

If I represent concepts as probabilistic-logical combinations of the salient attributes, that IS my numerical representation.


But, prob-logic combinations are not necessarily LINEAR combinations

So, to represent them using vectors and matrix multiplication, you need to use exponentially huge vectors whose components involve arbitrarily large conjunctions of elementary terms; and then you need to use matrices that represent tensorial linearizations of Boolean functions.
Yuck!  ;-)

There are two major problems arising in your numeric vector representation:

-- what are the relevant dimensions.  For analogical quadrature to
reduce to vector addition,
you need to make sure the relevant dimensions are PROBABILISTICALLY
INDEPENDENT
in the context in question.  But obviously, finding independent
dimensions may be very hard.

Well, I pose it in a different way (namely, what is the appropriate transform to get from space A to space B) but essentially, that's the hard problem. The only one-word answer is "Search." But note that a transform is itself a matrix (2 kinds, at the moment, simple matrix mult and the state-transisiton matrices that can simulate any system of ODEs). In other words, it's a point in a space itself, and such spaces are subject to the same observation/learning/adaptive processes as any other concept.

Yes, but if you need exponentially huge vectors whose components incorporate large conjunctions
of elementary terms, then your matrices are size

2^(2n)

so this is a really massive and nasty search problem!!!!

i.e.

n data points spawn...
2^n dimensional vectors which spawn...
2^(2n) dimensional vectors (the transformation matrices representing tensorially linearized logic functions)
.. etc.


Similarly, you're a fan of Moravec's grids in robotics, but modern
robotics uses probabilistic occupancy grids that work better ;-) ...

Huh? Moravec's were the original, Bayesian probability grids. There've been refinements, but the basic idea, including empirically derived sensor models, has been quite robust and remains in general use.

I refreshed my memory and you are right, of course. I read Moravec's robotics stuff long ago
and my LTM apparently is highly imperfect ;-p ...

There are decent arguments that interconnections btw cortical columns
are representing conditional
probabilities in many cases ... if so then focusing on the numeric
vectors of multiple neural activations,
rather than the conditional probabilities they represent, may be
misdirected...

Who said a vector of numbers couldn't be conditional probabilities? Once you get a step or two up from raw sensor data, the numbers in my vectors can represent virtually anyting. Even symbols.


What I don't see is how the vector representation resolves any of the hard issues in AI...

Your whole approach seems to bottom out on a search through an exponentially large space
of matrices.

But, obviously, reducing AI to search has been done in so many different ways...

I don't buy the idea that brute force search in this matrix space is gonna be good enough...

To convince me, you need to explain what clever trick is used to do pruning in the search space, in a way that is contrived to prune in the right way to make search work effectively FOR THE
PARTICULAR SORTS OF PROBLEMS THAT WILL HABITUALLY BE CONFRONTED
BY THE SYSTEM.

???

-- Ben





-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?list_id=303

Reply via email to