RE: [agi] Do the inference rules.. P.S.Edward,
Thanks for interesting info - but if I may press you once more. You talk of
different systems, but you don't give one specific example of the kind of
useful (& significant for AGI) inferences any of them can produce -as I do with
my cat example. I'd especially like to hear of one or more from Novamente, or
Copycat.
Can you think of a single analogy or metaphor, in addition, that is purely
symbolic?
I don't, and didn't, deny that logical thought is important. But it's only a
small part, I'm arguing of most productive, AGI-type reasoning.
Nor BTW are am I arguing at all against symbols, (you might care to look at
the "Picture Tree" thread I started a few months ago to better understand my
thinking here) - the brain (and any true AGI, I believe) uses symbols, outline
graphics [or Johnson's image schemata] and images in parallel, interdependently
and continuously, to reason about the world. (Note: "continuously." You seem to
think that some occasional sensory grounding of an AGI system here and there
will do. No, I'm arguing, it has to be, and is, continuous and applied to all
information and subjects).
What I am arguing against, rather than symbols, is what you might call the
"bookroom illusion" - which you saw graphically illustrated in John Rose's post
- the illusion that you can "learn about the world just from books" - or, to be
precise, that you can learn, and think about and build models of the world with
symbols/ text alone. It's an understandable illusion given that we often spend
hours apparently doing nothing but read text. But it is an illusion. The brain
does, and has to, continuously make sense (in images) of everything we read.And
if it can't then that text won't make sense - & it's a case of "I can't see
what you are talking about."
Edward:This is in response to Mike Tintner's 10/11/2007 7:53 PM post. My
response is in all-caps.
Vladimir: .and also why can't 3D world model be just described abstractly,
by
> presenting the intelligent agent with bunch of objects with attached
> properties and relations between them that preserve certain
> invariants? Spacial part of world model doesn't seem to be more
> complex than general problem of knowledge arrangement, when you have
> to keep track of all kinds of properties that should (and shouldn't)
> be derived for given scene.
>
Vladimir and Edward,
I didn't really address this idea essentially common to you both, properly.
The idea is that a network or framework of symbols/ symbolic concepts can
somehow be used to reason usefully and derive new knowledge about the
world - a network of classes and subclasses and relations between them, all
expressed symbolically. Cyc and Nars are examples.
OK let's try and set up a rough test of how fruitful such networks/ models
can be.
Take your Cyc or similar symbolic model, which presumably will have
something like "animal - mammals - humans - primates - cats etc " and
various relations to "move - jump - sit - stand " and then "jump -
on - objects" etc etc. A vast hierarchy and network of symbolic concepts,
which among other things tell us something about various animals and the
kinds of movements they can make.
Now ask that model in effect: "OK you know that the cat can sit and jump on
a mat. Now tell me what other items in a domestic room a cat can sit and
jump on. And create a scenario of a cat moving around a room."
I suspect that you will find that any purely symbolic system like Cyc will
be extremely limited in its capacity to deduce further knowledge about cats
or other animals and their movements with relation to a domestic room - and
may well have no power at all to create scenarios.
IT DEPENDS WHAT YOU MEAN BY "PURELY SYMBOLIC". IN THE PAST "SYMBOLIC"
GENERALLY REFERRED TO SYSTEMS WITHOUT MUCH GROUNDING, SO THE SYMBOLS HAD
RELATIVELY LITTLE SEMANTIC "MEANING." OFTEN SUCH SYSTEMS RELIED ON RELATIVELY
BRITTLE DEFINITIONS AND RULES OF INFERENCE.
THAT IS NOT THE APPROACH I ADVOCATE. (AND PLEASE DON'T HOLD CYC UP AS A GOOD
EXAMPLE OF THE APPROACH I ADVOCATE. THERE IS A WORLD OF DIFFERENCE BETWEEN A
RELATIVELY OLD-FASHIONED AI SYSTEM LIKE CYC AND A STATE OF THE ART AGI SYSTEM
LIKE NOVAMENTE.)
THE STATE OF THE ART AGI APPROACH I FAVOR IS BASED ON (1) MASSIVE AMOUNTS OF
EXPERIENCES OF SOME SORT TO PROVIDE GROUNDING TO SYMBOLS AND (2) FLEXIBLE RULES
FOR MATCHING, INSTANTIATING, GENERALIZATION, AND INFERENCING IN A
CONTEXT-SPECIFIC WAY FROM SUCH MASSIVE EXPERIENCE, SO AS TO ENABLE SOMETHING
APPROACHING -- AND ULTIMATELY SURPASSING -- HUMAN-LEVEL INTELLIGENCE.
BUT SUCH SYSTEMS WOULD BE COMPOSED ALMOST ENTIRELY OF SYMBOLS. EVEN THE
TYPES OF SYSTEMS YOU SEEM TO BE FAVORING WOULD BE COMPOSED OF SYMBOLS. BITS
AND BYTES ARE, AFTER ALL, SYMBOLS. SO PLEASE LET'S STOP KNOCKING SYMBOLS, PER
SE.
THE DISTINCTION SHOULD BE BETWEEN RELATIVELY NAKED SYMBOLS AND SYMBOL
GROUNDED IN NETWORKS OF MEANING - I.E., NETWORKS OF RELATIONSHIPS SUCH AS
SENSORY PATTERNS (YES, I LIKE YOU THINK SENSORY EXPERIENCE IS GENERALLY
IMPORTANT), ASSOCIATIONS, CONDITIONAL PROBABILITIES, TEMPORAL RELATIONS,
CAUSE-AND-EFFECTS, ATTRIBUTES, FUNCTIONS, GOALS, VALUES, IMPORTANCE WEIGHTINGS,
GENERALIZATIONS, SPECIALIZATIONS, AND BEHAVIORAL SCHEMAS, ALL IN THE CONTEXT
POWERFUL INFERENCING AND AUTOMATIC LEARNING.
OF COURSE AS I SAID IN A VERY RECENT POST, GROUNDING COMES IN ALL SORTS OF
DIFFERENT TYPES AND DEGREES. SO DIFFERENT TYPES AND DEGREES OF INTELLIGENCE
CAN BE DERIVED WITH DIFFERENT TYPES OF GROUNDING. EVEN IN A SYSTEM LIKE CYC OR
WORDNET A CONCEPT WOULD NORMALLY HAVE SOME DEGREE OF GROUNDING.
But you or I, with a visual/ sensory model of that cat and that room, will
be able to infer with reasonable success whether it can or can't jump, sit
and stand on every single object in that room - sofa, chair, bottle, radio,
cupboard etc etc. And we will also be able to make very complex assessments
about which parts of the objects it can or can't jump or stand on - which
parts of the sofa, for example - and assessments about which states of
objects, (well it couldn't jump or stand on a large Coke bottle if erect,
but maybe if the bottle were on its side, and almost certainly if it were a
jeroboam on its side). And I think you'll find that our capacity to draw
inferences - from our visual and sensory model - about cats and their
movements is virtually infinite.
And we will also be able to create a virtually infinite set of scenarios of
a cat moving in various ways from point to point around the room.
Reality check: what you guys are essentially advocating is logical systems
and logical reasoning for AGI's - now how many kinds of problems in the real
human world is logic actually used to solve? Not that many. Oh it's an
important part of much problemsolving but only a part. How much scientific
problemsolving depends seriously on logic? Is logic going to help you
understand and have ideas about genetics or how cells work, or the brain
works, or how and why wars start? Is it going to be much use for design
problems? Does it help in telling stories? .. keep on going through the vast
range of human and animal problemsolving (all of which remember are the ONLY
forms of [A]GI that actually work).
That's why I asked you: give me some examples of useful new knowledge or
analogies [especially analogies] that have been derived from logical systems
or logic, period (except about logic itself).
THIS TIME THE ANSWER DEPENDS ON WHAT YOU MEAN BY "LOGICAL." WIKIPEDIA'S
BROAD DEFINITION OF "LOGIC" IS: "THE STUDY OF THE PRINCIPLES AND CRITERIA OF
VALID INFERENCE AND DEMONSTRATION." THUS, THE TERM IS MUCH MORE BROAD THAN THE
BRITTLE FORMAL LOGICS THAT MUCH OF AI WAS HUNG UP ON FOR YEARS.
I AM NOT A BIG FAN OF TRADITIONAL FORMAL LOGIC. SINCE THE EARLY '70'S I HAVE
SAID "FORMAL LOGIC IS TO HUMAN THOUGH WHAT DRESSAGE IS THE MOTION OF HORSES --
EXCEPT IN ITS SIMPLEST FORMS IT IS TOTALLY UNNATURAL." COMMON SENSE NOTIONS,
SUCH AS "THE EXCEPTION THAT PROVES THE RULE" INDICATES THAT REASONING WITH
BINARY TRUTH VALUES IS BRAIN-DEAD IN MANY DOMAINS.
BUT MANY FORMS OF LOGICAL REASONING ARE MUCH MORE FLEXIBLE. BAYESIAN
INFERENCING, FOR EXAMPLE, IS A TYPE OF LOGIC BECAUSE IT IS A TYPE OF REASONING
DESPITE ITS LIMITATIONS HAS SHOWN ITSELF TO BE EXTREMELY VALUABLE. IT IS USED
IN MANY SUCCESSFUL COMMERCIAL PRODUCTS. BAYESIAN CLASSIFIERS, FOR EXAMPLE,
HAVE BEEN USED TO MAKE NEW SCIENTIFIC DISCOVERIES FROM VAST AMOUNTS OF SENSOR
DATA. SO IN FACT, SOME TYPES OF LOGIC ARE EXTREMELY VALUABLE AND DO HELP
SCIENTISTS SOLVE PROBLEMS.
FURTHERMORE, IF YOU HAVE READ DOUG HOFSTADTER'S COPYCAT, WHICH MAKES CONTEXT
SPECIFIC ANALOGIES, YOU REALIZE IT USES A FLEXIBLE SIMILARITY SYSTEM, CALLED
SLIPNET, THAT CAUSES SIMILARITY MEASURES TO BE TIGHTENED OR LOOSENED IN A
CONTEXT-DEPENDENT WAY. THIS ALLOWS COPYCAT TO HANDLED THE DIS-SIMILARITES IN
THE CORRESPONDING THINGS THAT ARE BEING COMPARED TO MAKE AN ANALOGY.
NARS OR A NARS-LIKE SYSTEM COULD EASILY BE USED TO REPLACE HOFSTADTER'S
SLIPNET, AND COULD ARGUABLY HAVE SIGNIFICANT ADVANTAGES OVER SLIPNET, SUCH AS
MAKING COPYCAT'S ANALOGY DRAWING PROGRAM MORE GENERALIY APPLICABLE TO A WORLD
KNOWLEDGE BASE. SO "LOGIC" OF THE TYPE FOUND IN NARS COULD ACTUALLY BE USEFUL
IN THE VERY FIELD OF DRAWING ANALOGIES THAT THE ABOVE TEXT IMPLIES IT IS
USELESS FOR.
IN RECENT YEARS THERE HAS BEEN A LOT OF WORK IN DESIGNING SYSTEMS THAT
AUTOMATICALLY LEARN APPROPRIATE PROBABILISTIC LOGICS. ONE OF THESE IS
NOVAMENTE'S PROBABALISTIC LOGIC NETWORKS, OR PLN, WHICH BEN GOERTZEL REFERRED
TO IN HIS POST OF 10/10/2007 4:45 AM ON THIS LIST. I DON'T YET KNOW HOW WELL
ANY OF THESE SYSTEMS WORK, BUT THEY HOLD THE PROMISE OF ALLOWING LOGIC TO
DELIVER ALL OF THE VERY THINGS YOU SAY LOGIC CANNOT DELIVER IN LARGE
WORLD-KNOWLEDGE-COMPUTING AGI'S.
SO, PLEASE LET'S STOP KNOCKING "LOGIC."
New knowledge - especially new science - comes primarily from new
observation of the world, not from logically working through old knowledge.
Artificial general intelligence - the ability to develop new, unprogrammed
solutions to problems - depends on sensory models and observations.
Let me be brutally challenging here : the reason you guys are attached to
purely symbolic models of the world is not because you have any real
evidence of their being productive (for AGI), but because they're what you
know how to do. Hence Vlad's "why can't 3D world model be just described
abstractly.." He doesn't know - he just hopes - that it can. Logically.
What you need here is not logic but - ahem - evidence {sensory stuff].
BRUTAL CHALLENGE ACCEPTED.
(AGAIN, "PURELY SYMBOLIC" COVERS ANY DIGITAL SYSTEM, EVEN THE TYPE YOU SEEM
TO FAVOR.)
ACTUALLY, EVER SINCE I DID MY READING LIST UNDER MINSKY IN 1969-70, MY
GUIDING PHILOSOPHY HAS BEEN THE GIST OF K-LINE THEORY - I.E., THAT ONE REASONS
ABOUT NEW SITUATIONS BY EVOKING MEMORIES OF PAST SIMILAR SITUATIONS. SO I HAVE
BEEN IN FAVOR OF "EXPERIENTIAL REASONING" FOR OVER 37 YEARS. AND I HAVE NEVER
BEEN A BIG FAN OF FORMAL LOGIC FOR THE REASONS STATED ABOVE.
BUT I SEEK TO AVOID BEING NARROW MINDED. I THINK THERE ARE MANY DIFFERENT
POSSIBLE TYPES AND DEGREES OF EXPERIENCE, THERE ARE MANY DIFFERENT WAYS IT CAN
BE REPRESENTED, ALTHOUGH SOME REPRESENTATIONS ARE MUCH MORE CAPABLE AND
EFFICIENT THAN OTHERS. THERE ARE MANY DIFFERENT DEGREES AND TYPES OF
INTELLIGENCE. NOT ALL AGI'S NEED VISUAL MODELS, OR EVEN SENSORY MODELS OF
PHYSICAL REALITY. AGI'S USED FOR SOME LIMITED DOMAINS MAY NOT EVEN NEED
MODEL'S OF 3-DIMENSIONAL PHYSICAL SPACE -- SUCH AS THE HYPOTHETICAL
PROGRAM-LEARNING AGI IN MY EARLIER POST OF TODAY. (ALTHOUGH IT WOULD ALMOST
CERTAINLY DEVELOP OR START WITH A GENERAL MODEL OF N-DIMENSIONAL SPACES.)
I BELIEVE THE CONCEPT OF TURING EQUIVALENCE SHOULD OPEN OUR MINDS TO THE FACT
THAT MOST THINGS IN COMPUTATION CAN BE DONE MANY DIFFERENT WAYS. ALTHOUGH SOME
WAYS ARE MUCH LESS EFFICIENT THAN OTHERS AS TO BE PRACTICALLY USELESS, AND
ALTHOUGH SOME WAYS MAY LACK ESSENTIAL CHARACTERISTICS THAT LIMIT EVEN THEIR
THEORETICAL CAPABILITIES.
AS MUCH AS YOU MAY KNOCK OLD FASHIONED AI SYSTEMS, THEY ACCOMPLISHED A HELL
OF A LOT WITH FLY-BRAIN LEVEL HARDWARE. THUS, RATHER THAN DISMISS THE TYPES OF
REPRESENTATIONS AND REASONING THEY USED AS USELESS, I WOULD SEEK TO UNDERSTAND
BOTH THEIR STRENGTHS AND WEAKNESSES. BEN GOERTZEL'S NOVAMENTE EMBRACES USING
THE EFFICIENCY OF SOME MORE NARROW FORMS OF AI IN DOMAINS OR TASKS WHERE THEY
ARE MORE EFFICIENT (SUCH AS LOW LEVEL VISION, OR FOR DIFFERENT TYPES OF MENTAL
FUNCTIONS), BUT HE SEEKS TO HAVE SUCH DIFFERENT AI'S RELATIVELY TIGHTLY
INTEGRATED, SUCH AS BY HAVING THE SYSTEM HAVE SELF AWARENESS OF THEIR
INDIVIDUAL CHARACTERISTICS. WITH SUCH SELF AWARENESS AN INTELLIGENT AGI MIGHT
WELL OPTIMIZE REPRESENTATIONS FOR DIFFERENT DOMAINS OR DIFFERENT LEVELS OF
ACCESS.
LIKE NOVAMENTE, I HAVE FAVORED A FORM OF REPRESENTATION WHICH IS MORE LIKE A
SEMANTIC NET. BUT ONE CAN REPRESENT A SET OF LOGICAL STATEMENTS IN SEMANTIC
NET FORM. I THINK WITH ENOUGH LOGICAL STATEMENTS IN A GENERAL, FLEXIBLE,
PROBABILISTIC LOGIC ONE SHOULD BE ABLE TO THEORETICALLY REPRESENT MOST FORMS OF
EXPERIENCE THAT ARE RELEVANT TO AN AGI -- INCLUDING THE VERY TYPE OF VISUAL
SENSORY MODELING YOU SEEM TO BE ADVOCATING.
Edward W. Porter
Porter & Associates
24 String Bridge S12
Exeter, NH 03833
(617) 494-1722
Fax (617) 494-1822
[EMAIL PROTECTED]
-----Original Message-----
From: Mike Tintner [mailto:[EMAIL PROTECTED]
Sent: Thursday, October 11, 2007 7:53 PM
To: [email protected]
Subject: Re: [agi] Do the inference rules.. P.S.
Vladimir: ..and also why can't 3D world model be just described abstractly,
by
> presenting the intelligent agent with bunch of objects with attached
> properties and relations between them that preserve certain
> invariants? Spacial part of world model doesn't seem to be more
> complex than general problem of knowledge arrangement, when you have
> to keep track of all kinds of properties that should (and shouldn't)
> be derived for given scene.
>
Vladimir and Edward,
I didn't really address this idea essentially common to you both, properly.
The idea is that a network or framework of symbols/ symbolic concepts can
somehow be used to reason usefully and derive new knowledge about the
world - a network of classes and subclasses and relations between them, all
expressed symbolically. Cyc and Nars are examples.
OK let's try and set up a rough test of how fruitful such networks/ models
can be.
Take your Cyc or similar symbolic model, which presumably will have
something like "animal - mammals - humans - primates - cats etc " and
various relations to "move - jump - sit - stand " and then "jump -
on - objects" etc etc. A vast hierarchy and network of symbolic concepts,
which among other things tell us something about various animals and the
kinds of movements they can make.
Now ask that model in effect: "OK you know that the cat can sit and jump on
a mat. Now tell me what other items in a domestic room a cat can sit and
jump on. And create a scenario of a cat moving around a room."
I suspect that you will find that any purely symbolic system like Cyc will
be extremely limited in its capacity to deduce further knowledge about cats
or other animals and their movements with relation to a domestic room - and
may well have no power at all to create scenarios.
But you or I, with a visual/ sensory model of that cat and that room, will
be able to infer with reasonable success whether it can or can't jump, sit
and stand on every single object in that room - sofa, chair, bottle, radio,
cupboard etc etc. And we will also be able to make very complex assessments
about which parts of the objects it can or can't jump or stand on - which
parts of the sofa, for example - and assessments about which states of
objects, (well it couldn't jump or stand on a large Coke bottle if erect,
but maybe if the bottle were on its side, and almost certainly if it were a
jeroboam on its side). And I think you'll find that our capacity to draw
inferences - from our visual and sensory model - about cats and their
movements is virtually infinite.
And we will also be able to create a virtually infinite set of scenarios of
a cat moving in various ways from point to point around the room.
Reality check: what you guys are essentially advocating is logical systems
and logical reasoning for AGI's - now how many kinds of problems in the real
human world is logic actually used to solve? Not that many. Oh it's an
important part of much problemsolving but only a part. How much scientific
problemsolving depends seriously on logic? Is logic going to help you
understand and have ideas about genetics or how cells work, or the brain
works, or how and why wars start? Is it going to be much use for design
problems? Does it help in telling stories? .. keep on going through the vast
range of human and animal problemsolving (all of which remember are the ONLY
forms of [A]GI that actually work).
That's why I asked you: give me some examples of useful new knowledge or
analogies [especially analogies] that have been derived from logical systems
or logic, period (except about logic itself).
New knowledge - especially new science - comes primarily from new
observation of the world, not from logically working through old knowledge.
Artificial general intelligence - the ability to develop new, unprogrammed
solutions to problems - depends on sensory models and observations.
Let me be brutally challenging here : the reason you guys are attached to
purely symbolic models of the world is not because you have any real
evidence of their being productive (for AGI), but because they're what you
know how to do. Hence Vlad's "why can't 3D world model be just described
abstractly.." He doesn't know - he just hopes - that it can. Logically.
What you need here is not logic but - ahem - evidence {sensory stuff].
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&
------------------------------------------------------------------------------
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&
------------------------------------------------------------------------------
No virus found in this incoming message.
Checked by AVG Free Edition.
Version: 7.5.488 / Virus Database: 269.14.7/1062 - Release Date: 10/10/2007
17:11
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=52718565-49d3db