the purpose of text is to convey something. It has to be interpreted. who
cares about predicting the next word if you can't interpret a single bit of
it.

On Tue, Jun 29, 2010 at 3:43 PM, David Jones <[email protected]> wrote:

> People do not predict the next words of text. We anticipate it, but when
> something different shows up, we accept it if it is *explanatory*. Using
> compression like algorithms though will never be able to do this type of
> explanatory reasoning, which is required to disambiguate text. It is
> certainly not sufficient for learning language, which is not at all about
> predicting text.
>
>
> On Tue, Jun 29, 2010 at 3:38 PM, Matt Mahoney <[email protected]>wrote:
>
>> Experiments in text compression show that text alone is sufficient for
>> learning to predict text.
>>
>> I realize that for a machine to pass the Turing test, it needs a visual
>> model of the world. Otherwise it would have a hard time with questions like
>> "what word in this ernai1 did I spell wrong"? Obviously the easiest way to
>> build a visual model is with vision, but it is not the only way.
>>
>>
>> -- Matt Mahoney, [email protected]
>>
>>
>> ------------------------------
>> *From:* David Jones <[email protected]>
>> *To:* agi <[email protected]>
>> *Sent:* Tue, June 29, 2010 3:22:33 PM
>>
>> *Subject:* Re: [agi] A Primary Distinction for an AGI
>>
>> I certainly agree that the techniques and explanation generating
>> algorithms for learning language are hard coded into our brain. But, those
>> techniques alone are not sufficient to learn language in the absence of
>> sensory perception or some other way of getting the data required.
>>
>> Dave
>>
>> On Tue, Jun 29, 2010 at 3:19 PM, Matt Mahoney <[email protected]>wrote:
>>
>>> David Jones wrote:
>>> >  The knowledge for interpreting language though should not be
>>> pre-programmed.
>>>
>>> I think that human brains are wired differently than other animals to
>>> make language learning easier. We have not been successful in training other
>>> primates to speak, even though they have all the right anatomy such as vocal
>>> chords, tongue, lips, etc. When primates have been taught sign language,
>>> they have not successfully mastered forming sentences.
>>>
>>>
>>> -- Matt Mahoney, [email protected]
>>>
>>>
>>> ------------------------------
>>> *From:* David Jones <[email protected]>
>>> *To:* agi <[email protected]>
>>> *Sent:* Tue, June 29, 2010 3:00:09 PM
>>>
>>> *Subject:* Re: [agi] A Primary Distinction for an AGI
>>>
>>> The point I was trying to make is that an approach that tries to
>>> interpret language just using language itself and without sufficient
>>> information or the means to realistically acquire that information, *should*
>>> fail.
>>>
>>> On the other hand, an approach that tries to interpret vision with
>>> minimal upfront knowledge needs *should* succeed because the knowledge
>>> required to automatically learn to interpret images is amenable to
>>> preprogramming. In addition, such knowledge must be pre-programmed. The
>>> knowledge for interpreting language though should not be pre-programmed.
>>>
>>> Dave
>>>
>>> On Tue, Jun 29, 2010 at 2:51 PM, Matt Mahoney <[email protected]>wrote:
>>>
>>>> David Jones wrote:
>>>> > I wish people understood this better.
>>>>
>>>> For example, animals can be intelligent even though they lack language
>>>> because they can see. True, but an AGI with language skills is more useful
>>>> than one without.
>>>>
>>>> And yes, I realize that language, vision, motor skills, hearing, and all
>>>> the other senses and outputs are tied together. Skills in any area make
>>>> learning the others easier.
>>>>
>>>>
>>>> -- Matt Mahoney, [email protected]
>>>>
>>>>
>>>>  ------------------------------
>>>> *From:* David Jones <[email protected]>
>>>> *To:* agi <[email protected]>
>>>> *Sent:* Tue, June 29, 2010 1:42:51 PM
>>>>
>>>> *Subject:* Re: [agi] A Primary Distinction for an AGI
>>>>
>>>> Mike,
>>>>
>>>> THIS is the flawed reasoning that causes people to ignore vision as the
>>>> right way to create AGI. And I've finally come up with a great way to show
>>>> you how wrong this reasoning is.
>>>>
>>>> I'll give you an extremely obvious argument that proves that vision
>>>> requires much less knowledge to interpret than language does. Let's say 
>>>> that
>>>> you have never been to egypt, you have never seen some particular movie
>>>> before.  But if you see the movie, an alien landscape, an alien world, a 
>>>> new
>>>> place or any such new visual experience, you can immediately interpret it 
>>>> in
>>>> terms of spacial, temporal, compositional and other relationships.
>>>>
>>>> Now, go to egypt and listen to them speak. Can you interpret it? Nope.
>>>> Why?! Because you don't have enough information. The language itself does
>>>> not contain any information to help you interpret it. We do not learn
>>>> language simply by listening. We learn based on evidence from how the
>>>> language is used and how it occurs in our daily lives. Without that
>>>> experience, you cannot interpret it.
>>>>
>>>> But with vision, you do not need extra knowledge to interpret a new
>>>> situation. You can recognize completely new objects without any training
>>>> except for simply observing them in their natural state.
>>>>
>>>> I wish people understood this better.
>>>>
>>>> Dave
>>>>
>>>> On Tue, Jun 29, 2010 at 12:51 PM, Mike Tintner <
>>>> [email protected]> wrote:
>>>>
>>>>>  Just off the cuff here - isn't the same true for vision? You can't
>>>>> learn vision from vision. Just as all NLP has no connection with the real
>>>>> world, and totally relies on the human programmer's knowledge of that 
>>>>> world.
>>>>>
>>>>>
>>>>> Your visual program actually relies totally on your visual "vocabulary"
>>>>> - not its own. That is the inevitable penalty of processing unreal signals
>>>>> on a computer screen which are not in fact connected to the real world any
>>>>> more than the verbal/letter signals involved in NLP are.
>>>>>
>>>>> What you need to do - what anyone in your situation with anything like
>>>>> your asprations needs to do - is to hook up with a roboticist. Everyone 
>>>>> here
>>>>> should be doing that.
>>>>>
>>>>>
>>>>>  *From:* David Jones <[email protected]>
>>>>> *Sent:* Tuesday, June 29, 2010 5:27 PM
>>>>> *To:* agi <[email protected]>
>>>>> *Subject:* Re: [agi] A Primary Distinction for an AGI
>>>>>
>>>>> You can't learn language from language without embedding way more
>>>>> knowledge than is reasonable. Language does not contain the information
>>>>> required for its interpretation. There is no *reason* to interpret the
>>>>> language into any of the infinite possible interpretaions. There is 
>>>>> nothing
>>>>> to explain but it requires explanatory reasoning to determine the correct
>>>>> real world interpretation
>>>>>
>>>>> On Jun 29, 2010 10:58 AM, "Matt Mahoney" <[email protected]> wrote:
>>>>>
>>>>>  David Jones wrote:
>>>>> > Natural language requires more than the words on the page in the real
>>>>> world. Of...
>>>>> Any knowledge that can be demonstrated over a text-only channel (as in
>>>>> the Turing test) can also be learned over a text-only channel.
>>>>>
>>>>>
>>>>>
>>>>> > Cyc also is trying to store knowledge about a super complicated world
>>>>> in simplistic forms and al...
>>>>> Cyc failed because it lacks natural language. The vast knowledge store
>>>>> of the internet is unintelligible to Cyc. The average person can't use it
>>>>> because they don't speak Cycl and because they have neither the ability 
>>>>> nor
>>>>> the patience to translate their implicit thoughts into augmented first 
>>>>> order
>>>>> logic. Cyc's approach was understandable when they started in 1984 when 
>>>>> they
>>>>> had neither the internet nor the vast computing power that is required to
>>>>> learn natural language from unlabeled examples like children do.
>>>>>
>>>>>
>>>>>
>>>>> > Vision and other sensory interpretaion, on the other hand, do not
>>>>> require more info because that...
>>>>> Without natural language, your system will fail too. You don't have
>>>>> enough computing power to learn language, much less the million times more
>>>>> computing power you need to learn to see.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> -- Matt Mahoney, [email protected]
>>>>>
>>>>>  ________________________________
>>>>> From: David Jones <[email protected]>
>>>>> To: agi <[email protected]...
>>>>> *Sent:* Mon, June 28, 2010 9:28:57 PM
>>>>>
>>>>>
>>>>> Subject: Re: [agi] A Primary Distinction for an AGI
>>>>>
>>>>>
>>>>> Natural language requires more than the words on the page in the real
>>>>> world. Of course that didn't ...
>>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>>> <http://www.listbox.com>
>>>>>
>>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>>> <http://www.listbox.com>
>>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>>> <http://www.listbox.com>
>>>>>
>>>>
>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>> <http://www.listbox.com>
>>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>
>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>> <http://www.listbox.com>
>>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>>> <https://www.listbox.com/member/archive/rss/303/> | 
>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>> <http://www.listbox.com>
>>>
>>
>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/> | 
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>    *agi* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/> | 
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>
>
>



-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com

Reply via email to