This is a very nice list of questions and makes a good framework for talking 
about the issues. Here are my opinions...

On Saturday 13 October 2007 11:29:16 am, Pei Wang wrote:

> *. When is a symbol "grounded"?

"Grounded" is not a good way of approaching what we're trying to get at, which 
is semantics. The term implies that meanings are inherent in words, and this 
obscures the fact that semantics are a property of systems of which words are 
only a part.
Example: is the symbol 2 grounded in my calculator? there's no pointer from 
the bit pattern to an actual pair of anything. However, when I type in 2+2 it 
tells me 4. There is a system implemented that is a semantic model of 
arithmetic, and 2 is connected into the system in such a way that I get the 
right answer when I use it. Is 2 grounded? meaningless question. Does the 
calculator have a limited but real semantics of arithmetic? Definitely.

> *. What is wrong in traditional "symbolic AI" on this topic?

These systems didn't come close to implementing a competent semantics of the 
parts of the world they were claimed to "understand".

> *. What is the "experience" needed for symbol grounding?

Experience per se isn't strictly necessary, but you have get the semantics 
from somewhere, and experience is a good source. The scientific method relies 
heavily on experience in the form of experiment to validate theories, for 
example.

> *. For the symbols in an AGI to be grounded, should the experience of
> the system be the same, or very similar, to human sensory experience?

No, as long as it can form coherent predictive models. On the other hand, some 
overlap may be necessary to use human language with much proficiency.

> *. Is vision necessary for symbol grounding in AGI?

No, but much of human modelling is based on spatial metaphors, and thus the 
communication issue is particularly salient.

> *. Is vision important in deciding the meaning of human concepts?

Many human concepts are colored with visual connotations, pun intended. You're 
clearly missing something if you don't have it; but I would guess that with 
only moderate exceptions, you could capture the essence without it.

> *. In that case, if an AGI has no vision, how can it still understand
> a human concept?

The same way it can understand anything: it has a model whose semantics match 
the semantics of the real domain.

> *. Can a blind person be intelligent?

Yes.

> *. How can a sensorless system like NARS have grounded symbol?

Forget "grounded". Can it *understand* things? Yes, if  it has a model whose 
semantics match the semantics of the real domain.

> *. If NARS always uses symbols differently from typical human usage,
> can we still consider it intelligent?

Certainly, if the symbols it uses for communication are close enough to the 
usages of whoever it's communicating with to be comprehensible. Internally it 
can use whatever symbols it wants any way it wants.

> *. Are you saying that vision has nothing to do with AGI?

Personally I think that vision is fairly important in a practical sense, 
because I think we'll get a lot of insights into what's going on in there 
when we try to unify the higher levels of the visual and natural language 
interpretive structures. And of course, vision will be of immense practical 
use in a robot.

But I think that once we do know what's going on, it will be possible to build 
a Turing-test-passing AI without vision.

Josh


-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=53256315-ae7a51

Reply via email to