Matthias,

You seem - correct me - to be going a long way round saying that words are different from concepts - they're just sound-and-letter labels for concepts, which have a very different form. And the processing of words/language is distinct from and relatively simple compared to the processing of the underlying concepts.

So take

THE CAT SAT ON THE MAT

or

THE MIND HAS ONLY CERTAIN PARTS WHICH ARE SENTIENT

or

THE US IS THE HOME OF THE FINANCIAL CRISIS

the words "c-a-t" or "m-i-n-d" or "U-S" or "f-i-n-a-n-c-i-a-l c-r-i-s-i-s" are distinct from the underlying concepts. The question is: What form do those concepts take? And what is happening in our minds (and what has to happen in any mind) when we process those concepts?

You talk of "patterns". What patterns, do you think, form the concept of "mind" that are engaged in thinking about sentence 2? Do you think that concepts like "mind" or "the US" might involve something much more complex still? "Models"? Or is that still way too simple? "Spaces"?

Equally, of course, we can say that each *sentence* above is not just a "verbal composition" but a "conceptual composition" - and the question then is what form does such a composition take? Do sentences form, say, a "pattern of patterns", or something like a "picture"? Or a "blending of spaces" ?

Or are concepts like *money*?

YOU CAN BUY A LOT WITH A MILLION DOLLARS

Does every concept function somewhat like money, e.g. "a million dollars" - something that we know can be cashed in, in an infinite variety of ways, but that we may not have to start "cashing in," (when processing), unless really called for - or only cash in so far?

P.S. BTW this is the sort of psycho-philosophical discussion that I would see as central to AGI, but that most of you don't want to talk about?





Matthias: What the computer makes with the data it receives depends on the information
of the transferred data, its internal algorithms and its internal data.
This is the same with humans and natural language.


Language understanding would be useful to teach the AGI with existing
knowledge already represented in natural language. But natural language
understanding suffers from the problem of ambiguities. These ambiguities can
be solved by having similar knowledge as humans have. But then you have a
recursive problem because first there has to be solved the problem to obtain
this knowledge.

Nature solves this problem with embodiment. Different people make similar
experiences since the laws of nature do not depend on space and time.
Therefore we all can imagine a dog which is angry. Since we have experienced
angry dogs but we haven't experienced angry trees we can resolve the
linguistic ambiguity of my former example and answer the question: Who was
angry?

The way to obtain knowledge with embodiment is hard and long even in virtual
worlds.
If the AGI shall understand natural language it would be necessary that it
makes similar experiences as humans make in the real world. But this would
need a very very sophisticated and rich virtual world. At least, there have
to be angry dogs in the virtual world ;-)

As I have already said I do not think the relation between utility of this
approach and the costs would be positive for first AGI.





-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34
Powered by Listbox: http://www.listbox.com

Reply via email to