----- Original Message ----
From: Mike Tintner <[EMAIL PROTECTED]>
To: [email protected]
Sent: Thursday, March 27, 2008 5:30:12 PM
Subject: Re: [agi] Microsoft Launches Singularity
DIV {
MARGIN:0px;}
Steve,
Some odd thoughts in reply. Thanks BTW for
article.
1. You don't seem to get what's implicit in the main point - you can't reliably
work out the sense of an enormous number of words by any kind of word lookup
whatsoever. How do you actually work out how to "handle the object" - the
slimy, slippery twisted ropey thing-y, or whatever? By looking at it. By
looking at images of it - either directly or by entertaining them mentally -
not consulting any kind of dictionary or word definitions at all. By imagining
what parts of the object to grip, and how to configure your hands to grip it.
Steve: Sorry that I missed that. But your clarifying issue is quite
interesting. Let me try to tease appart your scenario and explain how the
envisioned Texai system would process the command "handle the object". I
assume that you agree that an AGI designed to our mutual satisfaction should in
principle be able to process that particular command with at least the same
competence as a human. So the issue for me is to explain in brief how Texai
might do it.
First I assume that Texai has a body of commsense knowledge about, and skills
applicable to, the kinds of objects that can be handled. If not, then there
is a knowledge acquisition phase, and skill acquistion phase, that must be
completed beforehand.
Second, I assume that the linquistic concepts are expressed internally by the
system as symbolic terms. Many terms, for example objects that can be handled,
are grounded to the real world by an abstraction hierarchy. Descending down
this hierarchy, objects are represented less and less as symbols in logical
statements, and more and more as clustered feature vectors, and perhaps, at the
lowest levels, as no internal state at all - just sensors and actuators in
contact with the real world.
Thirdly, I distinguish between the understanding the command "handle the
object" and generating the behavior required to perform the command. I think
that you are conflating these two notions to make the scenario more difficult
that it otherwise would be. Perhaps as you know, Texai is a hierarchical
control system. I expect that skills will be present to handle various kinds
of objects, so for me the issue is to determine the correct skill to invoke in
order to perform the given command. As I explained in my previous post, Fluid
Construction Grammar does not determine semantics by word lookup, rather it
looks up constructions, which might be words, but often are not.
Given these assumptions of mine, your scenario suggests that the object to be
handled is one for which the system has no previous skill, or for which the
existing skill cannot be recognized as applicable to the given object. Because
I now building a bootstrap dialog system, that is motivated entirely by the
need to process novel situations, I am tempted to say that the system should
simply ask the user to teach it how to handle the novel object, or to ask if an
existing skill can be applied to the given object. However, lets move beyond
this approach, and I'll explain how the system uses existing perception and
planning skills to handle the given object.
By way of simplification, I'll assume your intent when asking the system to
"handle the object" means to pick it up with some physical actuator. And I'll
preface my explanation of this step by stating without proof that this task is
analogous to those already solved by state-of-the-art, urban, driverless cars,
e.g. "drive yourself to location X", where the driverless car has never been to
X. Rather than a futile attempt to explain all cases that come to mind, I'll
discuss a couple to give a flavor my approach.
Case 1 The system can sense that the novel object is not dangerous and cannot
be easily destroyed by its actuators. Then I propose that the first strategy
tried should be to pick it up in the most direct fashion, and compensate in
subsequent attempts for failure modes that resulted from from the earlier
attempts. This is like the pole balancing task that can be accomplished by
connectionist methods and no symbolic planning.
Case 2 The system senses that the actions to pick up the object are not subject
to experimentation, but must be performed correctly on the first attempt. For
this task, the system must observe all the object state that it can to remove
uncertainty. It must create a symbolic model of the object and its dynamics at
the right level of abstraction, and perform planning using symbolic
representions of its possible actions in order to create a trajectory that
satisfies the command to "handle the object". Then it must execute the plan,
repairing the plan as needed as problem state evolves that was not planned in
advance for (e.g. the object starts slipping from the system's grasp). At
lower abstraction levels, reactive behavior can substitute for planning (e.g.
when slippage is detected by a sensor, tighen the gripping actuator).
2. This discussion brings up an interesting
question. I suspect that there is a great deal of selectivity going into what
texts NLP chooses to process - and that they don't include how-to,
instructional
texts, like recipe books (and most educational texts), which tell you to
do things - like "take a cup," "add water etc" - and deal with a real world
situation, in-the-world. (?) If you're dealing more in historical
texts, - "the cat sat on the mat" etc - you don't have to confront the
open-ended nature of words, quite so violently. Hey, the cat did some kind of
sitting - as long as that's possible, who cares exactly what kind it was? But
if
you're a cool cat told to "sit" on a real mat that happens to be full of
objects
- , and you have to put those instructions into deeds rather than more words, -
you care, and words' open-endedness becomes apparent.
Steve: I agree with your insight. Much of NLU research is now focused on
either information / document retrievel, or machine translation. My main gripe
while at Cycorp was that Cyc, in the same fashion you describe, concentrated on
being taught facts and rules and then deductively answering queries. But what
could Cyc do beyond that? An AGI aspiring system should be capable of
representing skills (e.g. codelets or procedures), acquiring them by being
taught, and able to perform them as commanded, or on its own initiative. I
speculate that it will be easier to ground linguistic symbolic terms in the
rather precise world of computer programming and algorithms, but that truth
remains to be seen. (e.g. Texai, compile and run the unit tests for the program
that we wrote yesterday).
3. While philosophically, intellectually, most people dealing with this area
may expect words to have precise meanings, they know practically and
intuitively that this is impossible and work on the basis that words can have
different meanings according to who uses them - and that they themselves keep
shifting their usage of words. Philosophers, for example may argue
philosophically that words can and should have precise meanings and be treated
as true or false, but know in practice that pretty well all the major
words/concepts in philosophy, like "mind"/"consciousness"/"determinism" - have
multiple, indeed endless definitions. Or just think about AGI'ers and
"intelligence."
Steve: Actually at Cycorp, at one time we had dozens of Ph.D. philosophers
whose responsibity was to add precise symbolic concepts to the Cyc knowledge
base. The company likewise had a smaller staff of Ph.D. computational
linguists whose job was to interface NLP to the rather precise Cyc concepts.
My experiences at Cycorp with their parsers (i.e. Link Grammar, HPSG, Stanford
Parser & Charniak Parser) also have strongly influenced my choice to embrace
Fluid Construction Grammar. Despite the current lack of English coverage in
FCG, there is much less impedence mismatch between sytactic form and semantics.
IOW any general intelligence that wants to successfully use language must have
a metacognitive/ metalinguistic level of thought - where it asks explicitly,
as we do, "what does that word mean?"/"do I like that definition?" / "is it
reliable?" / "how should I use/order words?" / "what is the best kind of
diction when talking about this subject?".Life's complicated!
Steve: Given this statement, you might agree with my bootstrap English dialog
approach, in which metalinguistic skills are the first ones hard-coded.
P.S. If you haven't read, I recommend Lakoff's Case Study on "Over" at end of
"Women, Fire and Dangerous Things" - shows vast number of meanings and schemas
that can be attached to that word - and amplifies this discussion.
Steve: No I actually do not yet have this text by Lakoff, but I have some
recent experience with another preposition "on". In my first use case "the
book is on the table", I accomodate the following alternative interpretations
in order to test my design for disambiguation:
book - a bound book copybook - a sheath of paper, e.g. match bookis - has as an
attributeis - situation described ason - an operational device"on the table" -
subject to negotiation [ a multiword construction ]
on - located on the surface ofI hope you don't mind me using your issues to
explain how Texai should work.
-Steve
Stephen L. Reed
Artificial Intelligence Researcher
http://texai.org/blog
http://texai.org
3008 Oak Crest Ave.
Austin, Texas, USA 78704
512.791.7860
____________________________________________________________________________________
Looking for last minute shopping deals?
Find them fast with Yahoo! Search.
http://tools.search.yahoo.com/newsearch/category.php?category=shopping
-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
http://www.listbox.com/member/?member_id=8660244&id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com